var/home/core/zuul-output/0000755000175000017500000000000015140335255014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015140346204015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000306114615140346066020266 0ustar corecore6́ikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB >+YI_翪|mvşo#oVݏKf+ovpZj3C4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;FmZl8T*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) *ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`岮X7̝4?qKf, # qe䧤 ss]QzH.ad!rJBi`V +|i}}THW{y|*/BP3m3A- ZPmN^iL[NrrݝE)~QGGAj^3}wy/{47[q)&c(޸0"$5ڪҾη*t:%?vEmO5tqÜ3Cyu '~qlN?}|nLFR6f8yWxYd ;K44|CK4UQviYDZh$#*)e\W$IAT;s0Gp}=9ڠedۜ+EaH#QtDV:?7#w4r_۾8ZJ%PgS!][5ߜQZ݇~- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?}=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hA sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"8}pJ֊iTv)vtT̅Rhɇ ќuގ¢6}#LpFD58LQ LvqZDOF_[2aޙ-did˥]5]5᪩QJlyIPEQZȰ<'Xa>EE衢^}p/:F?}bi0>Oh%\x(bdF"F 'u Qx`j#(g6zƯRo(lџŤnE7^k(|(bƥYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\MA/Xp9VqNo}#ƓOފgv[r*hy| IϭR-$$m!-W'wTi:4F5^z3/[{1LK[2nM|[<\t=3^qOp4y}|B}yu}뚬"P.ԘBn방u<#< A Q(j%e1!gkqiP(-ʢ-b:dw>w3C=9k-{p>րd^T@eFZ#WWwYzK uK r؛6V L)auS6=`#(TO֙`mn Lv%7mSU@n_Vۀl9BIcSxlT![`[klzFض˪.l >7l@ΖLl gEj gWUDnr7AG;lU6ieabp៚U|,}S@t1:X _ .xI_7ve Z@7IX/C7@u BGڔE7M/k $q^hڧ};naU%~X!^C5Aw͢.@d!@dU}b? -ʏw |VvlK۴ymkiK% 0OFjT_kPW1mk%?\@R>XCl}b ,8; :.b9m]XaINE`!6uOhUuta^xN@˭d- T5 $4ذ:[a>֋&"_ }Oõϸ~rj uw\h~M il[ 2pCaOok.X0C?~[:^Pr򣏷y@/ڠ --i!M5mjozEƨ||Yt,=d#uЇ  l]չoݴmqV".lCqBѷ /![auPmpnEjus]2{2#b'$?T3{k>h+@]*pp桸]%nĴFԨlu |VXnq#r:kg_Q1,MNi˰ 7#`VCpᇽmpM+tWuk0 q /} 5 ¶]fXEj@5JcU_b@JS`wYmJ gEk2'0/> unKs^C6B WEt7M'#|kf1:X l]ABC {kanW{ 6 g`_w\|8Fjȡstuf%Plx3E#zmxfU S^ 3_`wRY}@ŹBz²?mК/mm}m"Gy4dl\)cb<>O0BďJrDd\TDFMEr~q#i}$y3.*j) qQa% |`bEۈ8S 95JͩA3SX~߃ʟ~㍖›f!OI1R~-6͘!?/Vvot4~6I@GNݖ-m[d<-l9fbn,'eO2sٟ+AWzw A<4 }w"*mj8{ P&Y#ErwHhL2cPr Wҭюky7aXt?2 'so fnHXx1o@0TmBLi0lhѦ* _9[3L`I,|J @xS}NEij]Qexx*lJF#+L@-ՑQz֬]")JC])"K{v@`<ۃ7|qk" L+Y*Ha)j~pu7ި!:E#s:ic.XC^wT/]n2'>^&pnapckL>2QQWo/ݻ<̍8)r`F!Woc0Xq0 R' eQ&Aѣzvw=e&".awfShWjÅD0JkBh]s9Ą|ק_;%X6Q@d 8&a)a.#ۿD> vfA{$g ăyd) SK?ɧ"0(HKkD4<80: M:'֥P!r "Lӓݰ@ 9n# " $fGgKQӦ4}Gn\^=-Y5PI dPN6 Ozځ/פ|5) F[ڣ$2*%&h v%9HN H~Q+oi?&۳)-nqK?2ސv/3,9ҮT9Cef˝49i.2DxatC<8iR/ƬйR֌vN8J"iJ. T>)qaY4ͬlyg "]BvW#99`TegõII kюHLa^c&/H^FFIu`2a$mc Ry+R:LڕDܓ>Y:]t.+|PT6=qWe0NƏw<6o3mv8k vGOfpEOkÈWȤMف lOc;SR&.w,qk>MPs+Xh4iyuGRd֞q鮺]m S{}]U kV0/ŜxtADx"Xh4|;XSxߵă@pE:y]/"(MCG`ʶϊGi+39#gNZYE:Qw9muB`9`LDhs4Ǩ9S`EkM{zB<˙ik; JD;;3!4 2Y.$Dwiu|+lO:k$]ԜYLUҞ6EmH>azʳ/A+ԀZk"f`.,ל{=wh|_qYj5M{K$gv>cDp"'0޽5xCNQ1G2})*'>fC۝'*)"5.E2IeD 2.ZdrN6Uœ=n8D-9޵JKw5ُJ,􋃓ZUꋼ0b1f87GՂ 1t_o}{Mr7KO0Ao-Y*Is\S:JzA(:i!eҎ\,f+,Ąt78~ڋ~?[F^.A'!,iGow3{'YToҝf5ޓ[he>=7S8DGZ@-#]f:Tm?L{F-8G#%.fM8Y='gیl0HڜHLK'Cw#)krWIk<1څ 9abHl:b3LjOq͂Ӥ=u8#E2;|z꽐vɀi^lUt␚ɓW%OVc8|*yI0U=nFGA`IC8p+C:!}Nh,mn>_MGiq'N~|z`|mu}r:"KiyGҪ$& hw#4qn?ܶХfm_Ov^ܶ[6j3ZN9t9ZMMM)I[Rχ/C|W䳮yI3MڼH9iEG&V 'x`u.̀ab7V<*EzfH{]:*6M x-v쳎M'.hO3p-IGh ܆hR ]zi2hB9'S_;I/d0oIU:m/~[*K1QA="D:V&f:{7N>^uU` c/Xo_ XC.l.;oX]}:>3K0R|WD\hnZm֏op};ԫ^(fL}0/E>ƥN7OQ.8[ʔh,Rt:p<0-ʁקiߟt[A3)i>3Z i򩸉*ΏlA" &:1;O]-wgϊ)hn&i'v"/ͤqr@8!̴G~7u5/>HB)iYBAXKL =Z@ >lN%hwiiUsIA8Y&=*2 5I bHb3Lh!ޒh7YJt*CyJÄFKKùMt}.l^]El>NK|//f&!B {&g\,}F)L b߀My6Õw7[{Gqzfz3_X !xJ8T<2!)^_ďǂ.\-d)Kl1헐Z1WMʜ5$)M1Lʳsw5ǫR^v|t$VȖA+Lܑ,҂+sM/ѭy)_ÕNvc*@k]ן;trȫpeoxӻo_nfz6ؘҊ?b*bj^Tc?m%3-$h`EbDC;.j0X1dR? ^}Ծե4NI ܓR{Omu/~+^K9>lIxpI"wS S 'MV+Z:H2d,P4J8 L72?og1>b$]ObsKx̊y`bE&>XYs䀚EƂ@K?n>lhTm' nܡvO+0fqf٠r,$/Zt-1-dė}2Or@3?]^ʧM <mBɃkQ }^an.Fg86}I h5&XӘ8,>b _ z>9!Z>gUŞ}xTL̵ F8ՅX/!gqwߑZȖF 3U>gCCY Hsc`% s8,A_R$קQM17h\EL#w@>omJ/ŵ_iݼGw eIJipFrO{uqy/]c 2ėi_e}L~5&lҬt񗽐0/λL[H* JzeMlTr &|R 2ӗh$cdk?vy̦7]Ạ8ph?z]W_MqKJ> QA^"nYG0_8`N 7{Puٽ/}3ymGqF8RŔ.MMWrO»HzC7ݴLLƓxxi2mW4*@`tF)Ċ+@@t޹na4p9/B@Dvܫs;/f֚Znϻ-MHVuV_K2k*`cKxuBG&24T}Lai 0Va(7K#ӊ!,ZDxFQO*lם>!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲\oUQ\Wߋ47WwߋKpwSSۘF,nC.\UߋoVEuY]^VW0R=<ު˜˻ x}[ێ'|;c^ M7 >5\-> m-8NJ\ALd!>_:h/NAC;?_ξqĎ6xMY(=ͯl~l8V0٨T zL{Ac:&$ ^CpH*DW\r2aR|=(L X1|wrO_g ux1^^V2޲jMi^b``Q#dBxV#NBk1;DAV$"*1]Y~ d->'I`ػ޶qfW|9_*[I] 8i۴9wY,Jl趺83dN/m}Ip87rdbM;J""<;dKdYFC)ZYpw/?T5(D0 UU!JƕȲ(Jnvrl!CUJ&!\,`R8XŨE7eUYq{\~*ϣB6IPvM|AjvE"Ҥ5rnY_øq{PE^״6sq_E'_{sU`}_l bİI&<}s#; |C=LK" ҽ5,p?"iQtr?&qk)GE3Y}uA J]8EM0M#YEۡO!oE<^KXo8<YMM~ T:(Gz[H#c$ϦΰՅ"w*A5_³*uo-Ongd<YKA>VIaňzbI3 dTqׅJWyp?6Rbch;x̻Re#c9-"u)]J|x iQʸF\%\p҉\ϊ̒I7Cnץr>#]&BA ;lMU&y g-Ȭ.EQ<}4|0Ky#`y- B1ۀDFb(AyT2 <E|)Q( `J2|59R2]bTO.(aBu%co:x4r4z=aTZTWդ}J,UKj]a/yEu2r³řjAZϖ6隔:1¦lκa Gf*ꑺ_p<3N%RJ(FQS7~^N_Y4 @:(G?b@b/(Śǩɇs–QTr<:k~&J_nuex3b%oF5%<2Bv̴־77?Qw`<_[=:zJ G6NQĝ G2hDR"N;Th{#u]$eoe &-L!Cj&sB@ػnG<8;S3ٓi7! ; x*Z bRrZÛeC ;gYz\ m=i `Wnܻ:߈)PP(ʹ֬Nle7l&4BetBY@m+$PoZ]TM˺I/k ({Sx{46\sO2}>o:>)yoOP`U}dOcw u1epR+ww>u~t.J.)!HoZ0jtAQwU KPӺ"=Ksx(0}Ssi8?y޶Y8 `7f`@ Qzyk[ aԴD]OMv57( 6XFxoJRi~"s\ [Z=i;\2\I"oZL?5N& b^o?r^]\6Ecs%t5v7j\ozA')"{tT4Gdm__&2:FӋA#DH>=EaDbӇsugws(Rd@;'BL_qjPBisy"%0+goF$mEm gp32`g nv0Fݽ0^+*+qvʿ2Rr GI?*Exno5;b a˼)(y SH P e2GmQP9뫓xw#J>>n1p=2 *t|d+ZFU"4 (2˳4o*C<>R>zAEvu.#PW$N8Y NUЃ+ҟQ$ )`ZT񑕚 z$yC)G'ɵmM53'3X(8Z6MCp#qbM2&`&W;" tu8(P* CƮG :Xk2}Ciдʆq 8muh{v3a[?>^F8Mֽ'p, ʀbWl]Lb0NqԨ7Ta-lC-Xj먵fx%qXt)~>!pmtVG#ZJٗ:3pWyywweq{րTܴq]b=ncenK~ˁ'A4{p趀B_HeUȕN(-(+&p lj2SkW!0B@Ոl!~32B\VCyb=qUC¯g¼vO6u Alfy@V_y+Qw4\(o XĵO$)݆*r6Q(#$1E^ @m\4.c"1uR{V޿HVu~: ׉S,ڒn2M)Ywc3SԲ},ώ b34ĵD^9|Cl }V4R3IUgVv("RW{Ҩo 0]k=EspkK\C\Wm-/%jGA&4ms *0ij[D+by4!K{*Bq}kU NbnKm]$j@"Tʶl4|8Wvl=QUS`)6n(W`UvKݏ0s AgGg(γmֽ\+Q9iwiH}Tk`quD6RMDfC!]/Yɪ62AZ1ǡR"-!yh(4diiˎQBDUa$)]2 H]m(4K,=BmZP /j5ctKllչK] IBS׹m)1x֘X goAKR}9wW6 u]Ǿ =}>^MQe #T핆h$>=ݦE-]KlvcԨU<GP׸jAb9mG⣷GӈVf~.&yP 66 ڊ⿫lliY‹"P7 =x䙞E<+6eOi8"tM;p.gewM Pt]AZ1SWRxQIHܒ eo6Wb&ظjl3mBPCF%kOHڑWS Q- V\MҵUzX6~d9VZ7 _\\fe0Qt>a.?P"&}ou=ܽQH;=mA~2q@Mq ݛjCl[C_<ĞU&2[2zyW}ǒ2ϱd^\'(G1Խ4>1@{7";߇p1cM,A/9_e<7b^A-/iwm9hfYqsliͽހۿit>Mp5Q޾ $]CgqB=xNOv8f"r9f39f /8q^'.u*'ۋ27'`MY־43j"fr # :d[M/`$9rq:K t 1 =w[Oe9PCtAb#¥grs(}~)"^ZI,8 b>?{8 !`wm[)t1ty0dIN4m[^߯dqI)Tz$N0NK#0ZwujXQB4ҍ[!DzP5 VeA{.[/^$QQQ*~m1Y,BJb49}D*Qt{>L!c8&UB&=>3yZ[alS׵4y*t\CjJ:?QZeUDNJ IT$ .fȚw'԰?+QY5叧/p TDxU6x-u-fϺ}F&ѭb:P9XEzgϣ>>tp:;?xzG9zK|"]rg \g@(8X@\6F4ynG,b՞D~{{GCMġ/ PqN p Z{,/h¼Eu%" 6R L\9(/ s#w:ދ[6]{nLAIM+00>|kf/ۮ309`x0 yo<ƺ]D( rlIMJ,q0_i0eTAZV\ZP)sdzxa̢ "kpq6T`Y{/#ދY:p=f7`+2\u (%=tjjl$a~:A߅{[ 8y 2M,2/axêF`9f!ݲ]ọ<:fGp#0$sD`MkJ M :w0CEʰ;%~祜Zf$p$Jfg6qU`Ugw O TrZ*);QH'(Ŏz\$No"OH Iz=N"ޜ26Av{۳mU,l^]qıy ^2C…x> mnѦ8bD~WWHn \ѯ_Mog9~%~ϋ&l ?|2d ?HCjK(1V8S5v9 _m{{X`uևm@b@AZ|զZ@-KOf{ɧ/6|LZ;d M vy9њs[v@ 7q 翤$/߄ 87brQ^ Ww%n>:,p0[t@ǷeoG$/3;2d5sRܼK t; ^*fo2`}ח $ IL ʩ6]OH;(| }U30D6 cGn:8ɩdCS 牘8݋qI3uS,e=Q{^s GÀRiye@m>Q2،sD5)> z< JIâY}i].fcTo w$.nyфO=FJ[vƄA2w >*S{2`pA1?Y= қF=7 #ÈٖE6ct-z_ j7M0{d9m3]L}&`/*M | ַ?$&ܰ襷(!KaC#`JRdYELG] nuC8P}._օkz4tj߱0.Ha0 Vb6X; YImgc/+;P c4(ml; ]vUž Y 0BYä9VVN+uDW]6clFfN`3|HqHeX(uSQC="9| (@<`j*2$ 8&mP kSJwSIOUi;4ڧ'wuVuv ٝVi:فPJ(ہP;iGv w N(uW uw ݝPi$݁PJ؁P;iG*v [%ہPowB=PoBUBw'# w 4X%4؁`wB<`K8Gg%,+=@fJG"[}m~DĖj+[4}6HL(Yxu:DcrͯŜė9^BJnj#^+q4Xr-1ocF;$+Oې||nwOBpL1iZ|*'0?Es Ytq+UOSpPVv w{6yˠh0?'{N@OLfUڧI˛hvgcy+@&m "Y,T6. {.\?q2MqcB+-US3GG6D;)giTL/"[oe}t~3@K xӋWbKɇ #PECEФ~f+Jt I*We,N1LOTZPCYT2kJON7'[a/WUԒֲ!?ܔRlâo2|e[a 8IezYҢᡷjϡoNGy8/r0xgDc[+.B<[@_mOtvN" Vb}44 $ 'U>f$xc(j. 퇴,Ϯ)<%/$ZD2P^]Zxl})*Hi:6f5Fyr%lgv v1w2I=]KJ@58<ϞWUWig '3EkCLǔț2Ij`&8|řcW{uCW%h~w}U)R=:Z}GGH/3biϺtQ(GI r.=R0$p @lSǠ*KgQDsr8BIY3k?%_-=PAИU;wXʐjQO-EfR欂-!hb D6BGD#@8\ߠZFst`8|9(Waԩ֦qI}&Bj0LoXЬ2i+u@?-釻Ԃjly%g=w^ gxAIx~Jo!],Yp$ف]ʫ~!mrz2"6jsIn)2i;8=7VZٗ"U.*2!sipa/I:\\%_-r9}#'rIB͢'X4sBFcm_Bk # m޷)sSWG+b.p]#VʓOb2A`F,!Fϖt%B,:W>.3@?+B . b7J.ͰhMX}V|JKT;i_LECs*Yݕb15s<|懄aP5K/w{a1"TK<&J'=-e"F=\8:Wdu͕Z]xW%+.*b,(/WdʁVs, 1 )$IMj8v d@Y3:I>å7?1񧾤N{t(ģ:584xNdW+$ )JͣʸՒ15PRB+V$5#0e1Wi+h`]{p6"&9k$icx\ltvxH$KG(ٛ0\&l)BHõY, gkIepr@R'pn |Nq8>Nϊ="wK@vdd{2%Z{:_j3bd/OE$- ߷q ӝ L'{ާwJoij?γmQI[`o)h# ˿.f?Tv_~Sjw?vۏ#R?g|7ӷl7Ϥ~2^q/}ۛo̿a4]yvI|\-W7hvzǴ$ҷoڶ*!8#q^[os77{4ɷ+^esb76ICK +wR!sU4^w7Za>a`prC^䖖p6wg<)S2?LRra"3b?֬UM&*g~ŵH ɈzhihbVym qYK%U(k^.W/Htc.8(yv|q ~J74)p| n>*WLBUx*o " hJ\QpUl?P@m>$=t 5!CƁb* * 0|-Z~HR Υ DKP9'W1P(& G]jy+OFxΔdaVKQ$AY۹G1G<@quG QF T<>T"ü:Q'tK3 j(+)1Suȇ\A0CI(eB3EL.ъ0r'%])wPg߅;`WIC!'x3So kTK]KBxYPS$i󱋃#l]JMJ1QĻb|t6rh ptۼ}那k VTA[C`cӢg$h䅳O.VĶx%TV CfX(Ƅو# R0ю=āqD`Z&\%2H: 8j6@sN 4H K[ƤY`"ERPEѥR苠`Q1HBIdR$)יڊp܈Ga$4v 932UBs͛_[s{Cqk{]cKzc]htm@`X( m #G}s}]#~뱛_N^$|kyQ QM*3S/]p99zD2Y_F/w==FŶ͚͌fRlI[N]pxN#^8;2V(ax38If'%HZ1&g:Mt=X{04pjWH%>8YqB94lỸq9FA̦צyNTUH:S/uXqhLa\m%(llB.zqU;eNد);eN/r7Ͼ~Tm_>-C 6 -鹼00S_+3iu?t$-XȨ0i9o=u7Z nY/nr]1r0w'9X"Zyf^efQ "Of-_*JO.oLl9.PPr 8m7*ͥ^'=_P89jn|ZsNxBpG|-Y1!&sŝw|bs|=Ej.Qgz!<׽7`KywQ'w[C96v@ܛe ȔZ'd[Jy!r^&BٳQG9m*;}97xC& v8&R^YSKI"̼)l|y Ƕ,hGRj)&eH o,tو8 x#&S$AZ aW_ 8)HnPQkƪN 1)x"?ERi=ipCy}l0Lu$E`Ks[T;+]ptΝu\$|!7 D~#t0ZkrT-֣r]Z._`tR;3%I`2.C}]pl?A%o@iE! `F:ô#Lնys|b6iuAr:b )4༖`K2#2P$u p3:p J#^-Efm~]L|] C&mx9ڔ0"f %[eji@׵ d^ o@-F' )p^'ED&EF/p^c)eK]ijG4d֕N )!Mnpkj#[*øU\[$fv͊+Ro Y7O]py/q;UUO%B0S\[, ZaĶmED g]R-Dw{Qjb*%Yo»x*o(WM<GRf?]oUE(}fw/ve 'C1]کԬNs}|G=+EМ s МfimR>Pd(bzg A'=fbJ h)t}|]pܶp/s 82up4fr׮ UZOQrop"tQ܅ROyA<.2x#˫ZGvPv!JϘlʶwK_$_)S!_n9w`RGp*=d)΄E$('hsbY.}.S??Ґ¨\:4#56ΤYDSٷ]Zҵ<NUHmaXhuH6, a#-]p -뒴H8z8w4w2'ĔwtH];(;3|0n| >F{$ٻmeNo{n0Y>~8)6:kٻMCJNd;r$Il~|>٧cMfG t{>5鹢K079Zty k] O$B2C0sIz MJZ}r{)ϲ,_&-kKwylo&3&RZv< BxP0~`*ap}YJ_6[q'y09M\\hBo47/pC0降ImS ] p$d2;&L@w񊉏20 Q}Gg0j ]3סxj͆ /'aWt+`;&޿3yLRW~#z f\O&~mr{3^^x<]G)81l6}XFx.d%/z'~T%K-7V~C0W~Rd&l)~޻NE$D!BWc4|Qc|\ T P/W#lb z8IvPu }y~Ch/ @g6tdEcUn1W2 /Mgd2TZsg3wm `hNo*z!nDJ?zha0?/I)M6s/W|bꖫ|`ͿkAF=~=Id;pˣqֲ2(7PQ0W$H׳*y`ih!ZS9 ezÖ,DxC+wqvOq3SSx8  .Fn?{ UKVzȸu9W"WTs$J-* . 9yS}3i߶aUU;)_I[gaRx6 ljo)Nd *5o.kk!Y)oǕuLfzy;d&'<=,WH8Lղpuq/ "xwQZQvp XlrוUMfuK8Ui忶6Us^w-C T6LJOvmvCܮαQ?ˆL$e8wo4j.l'eHf־KCsO߆O 3}_7V% IjZߛ<șE΂; ]RwBiu-ΚDiFR&q ka-qΗțH'PA7ը,y/\y/a,n[.G(G:!jG=Qg6dSm08eV+%MD|ޛPco_^n-! i?+w9c)07+ϨC"Rr+59Z,vj>2D̈́cO˧;Zs샰 t." dGE.zW)[_jء*#߲ԉTaBfe:zF֕~/?⠇:uR!||WtutmvR!cb\[{ab-wZCvkq^u1^>ٿMc|[cL F |C=`AϹRU}:"8=o+X;gtDa[Y<,*rDޑȺR=fȠ['&$/{gIJ,%PZNmvuJKԝ -YX.xʍ~U·4/*G'q,QV2viTZ̚AI 3aAcco߫&?,BWw qߴ d[Ⅵo- B#z-XƝv?zռA'@*h8m1'_z¡# /bck${7ؿ0,1/3MW%Ϝ6@,Uܖg-az,~XNy[szBcvP纎J"0An͟GlR",RZ7^c A9%yQFBLeBTib.Y()o|`}闗4'Y>tt6(0ngg@V]pPMc=7Fבƕ7 Nhƺf]34Oq!e  .EP 31 Q>`6Ռ5Ҍ7 ^5cLZ-Q$!GH9ȧ%ϸiF10 ,of4f4Adu$<3j<倓Gm x "< jc8ZHc`TXf4f4㒢ñkfDbckP$΅,6i H0rqJ[_4xZޤYE I7׌7K{9:Yqd3)tVb@zgo+9ҏ|vb5跥ŠD?۠ߒf`e}C6Z`C*iA>N!͕'h'ٍX??f`A_ LQ`74ڡi5c=)˱_|7ȯ)ޯ _A 5* C, L*jҥNl,Y*6z0ඤx0ǵ;J:+Nk:|釀wQ0޿ v1R2`RХ`wkDD8gaQ@LESx5YV(`yu*q^fH"#AF_ϯM&sq;# >%}Ď%c C[`%5Zqo4pq4'_o/"kWOhj޷0'_Yr|L |4jpw0"|vt2>B0&@674ޑm %\ЦFK't"f!m\ւ$\F?@8,ޗ"co`8oA?wvT3ct#s%:ٚ=Ɂ4g.spIP`wqd^ ")^U5>c$0_8sk!ϷqdԄC`BVF ԩo?'0xC ?[))㏞N`I9;ꑳ3[Ii1;QQt펵'ټcֲvvz hj҄eJ%m#|>_@Y}ICXQ &`C Mho>Y} 76CM>..2^|qpzq 4:+j3-R82T:SYzaZ6}3j|7_:_)7EK4OńFצes YDg)DIJ1c1FJt*Q;9>T̥Bte.r{'$RZ!;B ][`y"f&(s3eX,uw$2)Il442%)(y?芌E?/P]| q)˧a|ٺ҂?5D"Jj H+013d X30/C6>5@u\[f:v*u J9Lɩ)ɩxg)g7#ĞT:Xʨt2DS@h$(Pcéö)*TkS#/]&2F ;J (`FH3r`} kTC@g0 ]2q*u)Ś) \c'AJ:P,,D:Q .KnH.#tvd;ncP4R.(ĨKLC$mgL;L2ѩ:4q560)eLSu)?Ғ cN2#/K7 Jm/\9OhZj_6 ڿ]"ecvzz6S`ڧOk1v:-F3vts|CIq'v{IEj KmFLs[f-dFrU&s+Y=i"0RiJlRE&"a~&bK-$@) SO ``L, vJ2NK-1BI5XH 2|buoDE9b-5 ,CR$Vf inL{- =‡;n,aye3]m6+,}B /mmT\qr!ɹHэFTHiV5HJfFR$xht7X&Ae"jach-4VsIL((+ᾜ-`rc={%=B)ɦ|HKG89͂$ՠI0X&EHqR4c jq`J  #BDbQGBRpcƉ$X ?+ )U ߲)4>#p2lƑ2PK3b;b2;wp u֖DJ @+,@D"Gz7b+QVe&Z~)rFQ]İu19爈X"VLZfu2Yo!w.B+1,+ h*$ 57 . 7H/Ti0Sb0L p9R|Ɣ]&TiZP(RB"ׂ*'ۢ Syڿo,4vFAQQ yHx=Z/]4_]ߟPuiDiFW^Ùӑр=VL ̡CQb͇-:V/U&>pA)CN+V,1bJyPd$MxUZ _(5?@WمNlb%g{ U64|A5a[ 7 /GhsV0a1:'rVnPj/bAD] !,(0;\Au:A+0#7,ՀA1,n @K 7H40dHMH:$a&{9k)#18c_~P8w)BS&L΃Ս?> BIOvtl8eǕ/dc8,(V TP$+q<܊TpHb'T+"i$#!꘤>NS1+EJiUF(zê#S3Z(2OVHDm,2S1BJ"s*Q2n3ʤc,! 1Ǹ$ KZFpKcm'XGg TZ*JNPdQKb&D(HHI9dHE] n\Pv;NѐA& *=,n`X 9 +o[*/7:LS( ; 9n>n zKT \2 `p'/1AɁq"͇ #{*-c(hLܴ,6= Vhu{߹i4΂/LI9S`uЎl1n4g=ksƵmʹΟ%0h\W k: p 9Q6D+ c鹖# H mhjFsg%kP`` |JaZSivФ H[ٱ601v`pE5TC J\[HM9ŧ|Bq RLϺSiN+D LXcvp.sc"b<mB:{j pvpjqŗLL!& Yx>h|i_zZ+ m'?QqfZ_>=r#hL_|6v#' L1}0=ٶ+B0'2YW=/BY zqJ4Tǧ)@dꃾJ >Xa>~TH b%H/ocŌh0kė$ibcJ4ł\IÛÄCV}~,Kײd'v h~ u1'6CiVxA*ÏAthrerԨ GBڙ-2O=~%Bv%9q \+}pޅq.5Nrڶߥ ]Kߌ_ZD F mکe }ACua7 x-e iz3)Y!(,>A;ux;c_ N_n#^@He^B0nV]Eo[䄍v7/7ׅb9a?zu}>*pd'"2luF ªܦ./)(_(0vBw'Z<8?ٟjzgtldv3Du'k,ܪ:Sduo Kə'Y0E{Ѿ,l|)܇-*ƅ%%,7|Ё='Ch#[nn;tSEijU<YUtJdצLu2`.SyMֽwn:ЭVg 穉at5 +wޮڥ#{7b3;Wf9Yv;'@VIj}_Jo=70oUwƋ.V^Un=:ۣ)9CbcqHx@띎(ANh1n8B\m-)׃mIǓd))Q~ ހ؜F)fA횟?Mc;"%'(j]&2"0\En}'r*x-wUs}]-`4?TׂU~%n'ٜ! edyo%< o>-3';g^c1_?:QnM;IqZGVt{pr)}wܡ3MQԖ8'a3 gWf{(uf`A)sϴp즈CMnR}Wr-Bcø썸}D3v]-RQ|m8-aQ3,\eyҲU]U?6|pl290B6_ϿHm`Wdk9휯Wy]i/" 8xJWI`~҅{s G+/X=% XSZx&&D6ͬߨ4^>m5j+_'Gh9 F, F!Ԅ1s< Ͱp膨M^]*وZ^t]bN(*AE0 0D膮Mޫѵ%9]L(D84đMa&BN!zYeo|  ,,2>ÂygP*\(0J,yFI*h\fw';;!;:1,#4B&S,R~6yFfw|\8Kq0HX c6~ib$0F@jEQ(JZ=5W}>4colCN24%)xOd .TRPBRGcx T#0NOxdDFUu x8Wfi5 <1\Bf g܏j<=P3|]1Ir7&\ef4Q(ICd̛BFdU1P9|+^X*7][pa 񅆯n̄EJɈ+}L \KLybIDFH dNTgcYyhxQsZ1e^v \HGΑ2X% 51˔=g{C`14(ڶƓ$u@#]%ޮuDN" o,1@ջnC7E>7o !:wyZ>?T_)b)o'rwmM]w9Լv1;j͗/{Nrڭt(s]:ۇ|})_oy}VtLt-1>Yoﵕ7i{[_/_~ԧ+ޗ~Y-ջ?4'=Ӈ96MyWvX ;ax>>!>z8/"W_} ~_uozYwxl7^iqIB S3ZJMI]3 㥂yppg^jݣ . )et$$ x0~)qK^=@8򹺕baH z%-glQ/N\r z%e 8Cb|R"?}8{pb>CǮ>򗬦2qkji{ZwaQ !---DKq1Qg&'C8xGKt*U9 \0j+HZ[*񘭺1޲PV>H)R^/j+*wހO _M0j[.I>#[wd7`P+zQT; hʔr MIY3xUz?^6D_{Ջ_(bջ׼B*\OWqHqߜ DrّǗcD-rkW9xvg[Qd 'ZYm dDF!x 1L[,sB 55<@Tײ2==OG*r֚%<2⋄\JGcx T#rmIsbˏuѷ!y?ԀQW1:I)Ct&!CԪt0yQ2V#X m5|+f:DEd/$9p'z  DFdU1P9|ʵkZ̓lK D'X8?V}W/ \ݺzY0gc48aܰ邽okPb%O`.DhŴ4pqn.2 ǰ"rQ/hޗ sڐiG,pG|L40ȁg?Fxf8F"Ҋ%v}'L,H 슰$|RH9," .K %&|ǫC>W嶞q}mn>N஧$ObYK)w8|m*M/kծL[r<|ry)%r)j9"505dcIuȔ"pOϒYyhFU?5 k[} 3r}ƽakY;P\7mKsD c$.o8+yŹ> N*;cr+vJS|¿7~k hiPFvR֡QOl !Ґ- >BFdPa*E>WPP&&C gPrbn3br(Z`%h9S,kGθ0}0}(j!^z~ );Ѳ͆g1$}ǫ\xx1]b'uDtbLFd> 'AU^`iJz7r#ZZW7/UWP,p¥ C/r?03$9СCNdPcn5J/W~}\RYR2KODpF]=D%E>SZK=l!$uxeIyeR_x*cz|V͢ w񞄍ߡ/Fd\1'옡 q=u5wm05=g0b6P&γ +>RrI@‹`0lEUiJ3 b %65dAc\[V3VL⪰P΋6xb8‹tyg(*Vm&xpf$E+\pXإm&5Z\՚zD?C삾Ⱦ^hGYK80̙@` o`eВ֕I1a"ڊ)( { -] OJUKM( b*A ;MKEr(Hz9Zyn <3|ʵn@wEN +"+I!NIP*Mx鋇w{ G_e0s[tLL Qq) υb|e6b|ɸUГ 2’hm8g8.ͣgQ~&s#x {QsU^1y:r0r.O[LJ2 A{n @2v 7C"3zQ}-^.w%yNp<9 `{&2~?Dhω4%T(8#zxh"Ԛ8FEۣ--6m2{ލ%A'5diwon& a`wFd4 cYN屄82"#+'@yo-+]#ЭA ԝ:%~* fY ͗L,|]ۣn1Ot#JGn&lU0uYH\wKlFUe]77y{1 :&AS&<@·Y`Ki8&e}&03W0b~d0f\VkOٱmxry^G~9e-Fdl< mK x@hh& ZLܙF|nR{IVáR]?},$WQlɈoCm(*_0YP hA,mJ[>8ZhPh+9Z{s5c+6)?y.lcuBT{HXv!os<"*NBBA qK>txݍ7֯G2(EqbFAFA~^W9oww7N7Rg)풼vTSRq*;M8K HFdXC,F\:uܮ_W z"س- H6!#0HKS4$m!$&E` V<5aE>S.l"v?#Fd4ht{b鹜{P`ˡ㈌ -Xv"%uw_8 MBy+3桡"Tm›/LQH//S yAdr~_vnE~&7p_!A"lneڢ >y2s\6DE {`CC`C\%ۦOQ^fSՄFVAp)1)!EK pV􈌾cˀ"rRۤ j͈9kxd-,%w |+Њii"}Έr+0S ?"yk+Cѽrn vLr i+zA6OL `v^5G6`rF}xa#^Fo<E>WrVNؙ&zO̤L aJKpȈ c"r0RLFauj ֻ{Z=eY䝉]W:Ɨ"yk5M\'8yN1:Nx7kzD߆ITws"ԶaI3Vx.)2xf'Tω`xQs\q&_]?l7D$k :-?j_q+ffhA?H=@KԦQ@;QO#fnKf (%˫uˆ8a QQ"{^@(uuy )!Dkt^2ѡ+"] ԏx< 7"a>'+攺\.}(]IzI@i~yod{eEdd5z2CuwoE>W5j%gpQ\q>9dc5j[\.Vy3Cuh9)q)|0 8#2Fq`ꍆ>$|(,g"JL-䲈>PA*( KHL9U#2djN/~ 6Du\ [lƴ]SwY-vֳ_|XeR>nNȢcBZNw-j @iٳIޛn^xgOl.ɥ1 Avˠ.<>0^Ia4~kt9vNV6$t[cX*3Vp<(aVE>WBTGm+- \SE@Tǥ"`M֪:PUֱ+e5 cȶĽymWFN8Х .9$' ND[2H|qϒh2z314(*8m8u$Uپ0b5-|ۖzB1SnA[SJ3ء0a="VhPyŞkbzww=cm"j%p9<9Psu^T™j!⯫ՄRZW( w*T"E8*>mnH yGr|xa1]÷#j{_6${ȇd? DY)ْڶMsKU$eWhݺbVVagv)ۉb~DUG*d3ww%u&,=)}bJgsʨdn#`&Pm=&S\mi:5\vUMO W,stl⺎fB)K1iFq*D7/V"R 1~An{1Zs?+&^ 8N= ~2y2K][-w %M3eOr:(gʳ8 rYE@!a[&w: |(oρN{kZKZ=Xb4=&`InɱQ@J0*L݊n{-u})2)ilws牦#ꕰ;1%6I]kē 7ʇ:TT3z\|#A8EY7{̫Ǣ 㳓׉`־_Hi׉"T07uuΈ 4! #ol >KɥՍ8,ʾ)|V_GN cr-0(-wF i v}"N# >1["Ei=dhCWC^'tNA 946FId$e3"v4v_@ %<Ν|=-DmWl,nd[Ϟ\:dyiܻiֆs5=8FucQl1$us}蕫oQZY\(v+ ȁXC}F _c>(Ӄhu+ hXC}GphqhL8HRj21&3lVcN[@ă85|uk^NI+.A"8_S[m{4hRFة hQnu%vE}MU L']=PE%8N(O!S UL&l18d=:!ˇ7ț{ei+,OJQ9IjUn\,!XX>Wa}Fz!xgCƉR/4EuJ5F"$ ?hhuhBGci~WGI9mow1 n5"(B3a쑬b"®{(zW츺 = n~- !w1"eysbPAa#B ^xpl85fɁއ.aI3D$4gʙ"ΜOi | *5*Z!;}g׿'U6}`vP)!`w`nga~ cah|!LAB/R)t1CTH-x h0e-IQۈ S+L>]Yn,ė,} hѯbGC> x> wy8[,եc襙Th>N}a d*6p Phb[a{\lI0$C)5ˡekxt hXbR&$jxlExPkC7  *B08fi ҃_&!Qq P߭?OWS]s1xb1pcӫFmWB6b(k-|f?s}W@1g=j#۫n۫Ut7IN,r(oGn]~Gdj#'ulH^?%m-:xb21B@Q*$= rVO:JNRҸm08ywֻ!.eI#A2r%֌2dI&81OjaJ^g2x@;ZQIxDޓ{xv$zzO>yq5/OOz@cG!zze]0 sACmQU8zW8 "8)1!(Aʙ-4Kvޡ!ЌL"S:)Ⱥ%Y|BTe&UЧ-Ӈ; 7Gku.{lb~$ s@A.r -Iqi>P¡2_"K|UGs5̅%*%5Hm-?b*5upͷ[QNBj8.o9|[O? :$^@|=[A!AI 3Y&$ ^f2Id'=c 1Q/C>"ǹ LR] bUSCP|1V-#猇!'["ǵD"<[8 Pۆ! J_dꝶPnu\ػiȃ'0c"&q"GW:׿n@]b11`9Ϊjh+,Ll2ojy~K_C<n1F0MtbYI[f,Z΍B XU"IR3]U2*_{wws%7ˉ?Xnd:A69YPRr̰C-Vz?M]KRꇩ}p3槅  ?jܶ _6~%̊JOee נNX3DR78ca/W_Px9]y_zEYv51qҶ?Kn.>DӤ7?VD9+vfU}Sxi~nMmRkU,j7l-*3~-w6#,3nWǽ5gq{;X*Cus:Fsb$UGRRܴS 8;/ }e짉J'Ťt軪XN;q}{oߖSS{6|m ="W6O>]'k{eQpߊS}}p@O?Uн;8\!qlSOr|OzRުbެ#؈  auk={ƿOŴyj\|\GKg5ڶhn Sy-1g~5C g4ӼݽξugWi2'šh3='fZ6u+[y|۝[Y7q<++}Uۛ47ܢ>[V&q2NEGU6Z8S/gYb\s„BBl`;4DŽif+wNo!ւ_Z:}*~m58E8^`\\BcYr}|Ae=Xk;&5 ܨ7(0gZ%=u>cJqcќMHl@_^6یble6Q*(M։)Mboc$|8",^uq|bpXf-(PÇ}7pv2THl`5 lyhu(yiSs-G0S p_Sp"2xкI°7#צiMOG4ɧ?/wGz} םoC*c( N4/7;7Ss# XѐBrFIeD .F'Vlrϕ@|ņT Brݔf>c6,˳+-KlS]Y4Y <\hX=xC6TL9/b=psQ%HP)<_- Xػn#0y;am#9-Z%uߗ|OǓ$Q0e^nnۼĢnܿp*lCe7 (I9w< w/eۇх;=RoYOF f2 2igy?/9%ôKKbZILO2ONajI,=I BI8ZdJ,Q-laO4q{hQ)bO&|5{r=yۓ͝75R5Y,/Y]+BWr.~2t턷Fnaet7Sj14I7@(,2Ci$%Ѝ?6cnxϣPYIcmRqR8qgϡ*exRΠZ(W9$+4!B`:Jau([; ~kq*<-2.&C׿'m8 ㊡TYgph;RɁ80`6zm $a%Isw1p..B[Fv DWDK{n!ԥ}dE1+ TP,AaS,3i}C|`9HQ;y4LS8pm0!i*E-:Fn򊝷wu9uj Uܿ)0x!gyK=~!5]Fv9A0Pr1lq夔RGSfs^d0e h=N8JܜĽ:kY࿧3#5! qRA!(o(88Km)fpk#r>`r8/PRl0~VO ]M~4?̪OA+;Jſ= \.m|o0JՁ]{HTƋ\\:.r%o4F7 k}0VOѰn_ƶh!3lZ֨X xfw|ĵL~ȝٻ_Nfm,w|:SM3Oiu490M2g=< 鹶 HuTӛo& j7k. ͞E/{̀8,*)xtzƱLz0Hbsk}+xjW6FjećqĆg=JPrYTٷNSC.̎{FsV5!M5aﳐiÈG>'4(%E]-| !E;@|BtRߍ'YrÎ6} FfHTusV3|8{N#C|sSu(¬jEXuaM](n0:х,*#E }ORԌROx0M0C) >|rR9sRu@.w x1SCQBUoc)n( BfeєТQhJ>^ANd %..ۇ}X'!/,ey{;xnNeo!ŝatqh Y"YD'["垄D(/VJ%?ea A9 y,[tRXfH9so: )xH?YUmb ?F].(+}=v]'~:Zv8yC y( + ,9W%/6XN EP+恡XpQdA)'VBh(ռhmuCXX4CwUKxsh Y24ߦS<3˨7CgFVVS`Hō딦*C{C qʔΙkTwΌv 3FM%C2r69s/NSM]3E4=%Eש?w߁ayxue\G8nw#PT rM"G}kE@FܳX*)i \@0 0b>Ff\1}}1g3!h=n8* ^ F P Jv %a@ ,xTX?}>N=hNKfNhFRWUIgƁh720L@N R  &ꬷ/ =6 ❥Q\#+\ ɲ2^kVA.)l .J >fo=`~oۇхu%+E0fz i(8HJ]EsYķE% FgyU ݬ]ۉ`6NsQHǫ͊cT OiOHG M([rT6|05t:/Xewlne)oIatr(ŁRP-Pb$L9)tlGN+ 8+9өy>A Oo>-d8۠rhT[p/Y+)C/w{`?# H0d mkFO, \QԸ+*dArƢO8"+7 U@Q,Jǭ=I_/%f5@QD=nP)*13AYTLcA TH4E7(3(8 ~sv;18JOPF&@(Kk"))E&NF:箓E4 v; ;_{^hČV18W#D# VR0\D+`tؕͱH $=wo͹ǒY;9\Lg Z ٛnyjgˠ?y[ĝlQt)!\4r3EwgQ"i!TI֩JGEX`@vS-3_zo(Ũr&(bjhT-(+ApFI-OU1{<Ŝ]9.ٕE|m7~(a_:@h:2<:v*'Ff<VN0d(t@ FkJ@z/덓D0iIDBjTPy4 xW)ǡtYjN1Xib)όg_yL5 bya}CUM:Pkч{U[a.ʕA#p*j[B@Rŀq",(4`˽ܠ[}\❮:(m]b=Uv}f ts}%f5X1 a1#P*P Q2E)) "N޲}εP77WnIGC ƍvS`fO NZ,g+zf%<\ZeNٙvt9Y(Ɲ8=>5Q2xcOw5vfbwu{xu#BiҝmtSXtS@B%7S@ԡjNHe"$WQʞ Tih+eBZ19ke؇d+Hq( @i)Sn@NGFۛ(ӋULjbcg~Pz]|mZ_H+d.C)!i P#(z%q\顕k6% ~-5YoGSaxд8Kg˹͆\-u%v x3o L>_dZPa|0۝7aS&rCIs|-֗PH2, y^1QƏ~yw<{ǖĿ*(u*go32<1߲=ꋃ/pBrPNzW^r:8OӋH9GKh {pYAE!bR!h>D9RZڿHuM34YBvqʼnҋ, r'ן جyWS3i$G;P2TCY]yN#auHky,֫>ws9A 7 +0 L߽8i˨?"&:W_ 3GDT".ppU 4Q ..A3Za$E Wx'wEKG-"A(NcwmmX Ќy-E?vc:;=V:J;{XeR9,,%az G*T}uH~Jo1RQ)H;׎q X?"hp "psEU-5j"羙hm*JAyEqdTYċ6]xf`LPNa`b,^#ëm=!BicHqH Üf'5xL'5BZm"<͂@qY2 #hhlEF+C~ tٮԽ$\ %_Bzʠv!\ҩP*,vW'z/;8"HԷL#+5D6D/b"`7Vhzd x+qRa9BM/NHکN`TweLdQJfMJMK[K#0826vJhq24UOL钄-;Q;<(s!yzR1=6( {J`cF\`FDTuIiIq=#~8TI;Xj;aO <[ )ɻ,zR\z? &zd'C"U#m%]],ZlLsR70.ǑGqpE<ԩ|<'x5"!8k!GUkfld\fhͳ!x8KJ]N2v |pfьx+0ZIڤ-wLd 8v>`d(u?R|֏* Ȯn.}( %'H:e2F`#C\qy@1+*Y2t.ڹ_KzN8eƇluݠF_?n3gMʉM=2 Sk~<ˮM^/( Nnۇv@~YR$mV1suRg&Fנ&v đ9Q r KR3gF)o^I[c~ >dz&',}9IS( N5~aR8eHGb{ԃ OCLh4I1<ˀl4s O1D7:&0,A\;^qBR;'2cR?MP!f5W ԇM72 ^i jPYQb9,:8_05!_st^5efSy1aIb4'ل#=~cm(M.Vl㗜8rh~k6Y_~37QIa dH%iaoҵT#GɀDQD$ϵQͿ.Srß] NJ*U3!30#Sw 9I43w _h5e8i+08KO9$bO֝DGNMi!9)\Kr2HF{Kr:j$0$zdG`|L$2v&3aTÀ:;y T}W6*A8cVUYNpbδfpa?l$ na\ɛûċv|or:|ۃ&qspK`Y]~m2l^2~.&7m^ؽ2Mxմnk.U kg0CԚ2kLB| Y]kmOS#`R1m\cb%6wswf䇑ekx?5>4q_ӨǛ] ;?/_ޠ+^mșﻤ ^K\ֳuO|ffվo|}~z0 ŴBU^W7~Of /t hoClGiW C{0o(}ϕiNs6 +Cۍqm~+2Єk@>\Ĕʀ .k SJc01f6HN7gVy]*%jnv"xanp=Gp7tz6;r_)m=&E5۫0p]dn wv2yr y=)ϟ}w9li?6)j/zFVi zۗWpk`>mmA߳r{_Q:v" 7h-Z]<%X]ps`~|̻{\ܺYMoŸ?~iۥy;@!~~uf]:m=/laD #0f6zNn d;3`~ޏz,owƱv8&?@Wm\ T?NK.`mai] ^Łf-'/}g /*9y QGd1Vz\)vt6~:{p`JI+-WtllEmyot(IrѥP]fND@85jTSAʠ7e*Woy&1!&k!aSMs`N5f^ ZIVM&9Ըo{ңy]߭]AK=zbma7vIt)+]7Zxr'kⲭs^:F؁+ @!->}<|[>{EFėF"yUg1Wj7U\S/w2h;{}TnmKT8UӰ6Vgٙfgivvfgi6>:N4;;N4,Oզ623P > KaJ1  h ?mϾ ~T^~I~Ƅ 1a&x\ a㢬kNQ%Z)jjSYIjzggO+w>γ=y>i3ov52@ܧڞ<8\~|A>ӱs J; k5DDqىY@;ԓϽ ׋9r lHr%6")l]),*UYYq'lYذw Jo=,`*pGYљMyX6ݘ˒c 9%2_U)ilC͈>yajBuE (rXSc=S7<@=G/1@Vfzț矍;8#nrﲩD`]MH!&0nTetmzJIP;lr5+Wfhw_nYZ-[BDmz6/c晡̛;-I2{ͥ:ޙLPtfℑc?_]txj݄yss.T˂D= cd+kccq8Eg[`Gt#do la/|wg>ntz8 a`X7. !*^8j,GR4"ΰGZ CM9{˿܁._.|plv=;&/sE5 E2*$VPSc8c6w.Zy-׹Wa$Jʫ7JQ}@*N*NuGm ѯe5X?ba]aزЛ/W톲Jo6k܄IucR7ecavwd#x>&ָq<| ^Ն߫q ٜ&^9Y~99RD˱+'X3jDCP$mz_5m_e.$UOie˔Ny 1l{OJᯬeF$Sr$dut)<&jdn(|~v3vif*6*?AŠoD Q'ZEԨz(^x] u.\f]0Ǜ_+^ƥr&דa$bRby뢪m*| ƪ·ժL1Vp+Nn@DB.[cdIo Ame>+Mk򃑍U؉^emF S</r2rLأ'tH 8Xۘwn{kҡVX^xW `zERC a@-s8m.GNtt9BIļ𤇮Ϩ܆Jg |umW_VUUK2AFJ(d T485X q)1Uc-=vɟ/I:OnZ:$ A>ُ?tr (ja la5 "<7׫wG1gulR|xܫ t!3?V\~,ǙxIxOC}~]W/wXG3zϫyͪxum1o ݾyWocyE͝ts]wG19o6}]OCUC@Y$@_?qb`ǟqNIxGo,Y;]JSgY;uNSgY!}$Kxb*@ uЃѨ B8%JȌE}Q= \ûx,"ʙ$vV>fp@u|R>daϛ yz"Ϸeڻi|{._&/йs9F=@hsV ;Ja@Py$Vj5bK6ѥ&O$6e0.uFH! q Lau hA9d՜aڐX7`ⱬ_V/󃶺Zg7W7bn9HO}i۪W ZeBX! 0rFV9+'; BJQ6tR] kFqmOsү7v!&f){S|T5( qÈfr$--!="ɼq.6 ´S4SƕтB`j8L! P`ǃ2Y C Y0|㪕U[.cI.HǤ0N@u`(=[N68؉Vp8$&&c;IӒ%MRza(E)޵+BSwWnQ^<ϻbV<M;v;)v;Njv;x0DTtj8w<Ѱ3 b}BkEN^ 6NL)ҕo:.ѝ$[?{ZY9fD{ W/&4]~ 7:8G &5¿Guxr&]>߷9, OETȻ5Ãϫa4~yvvu՚|5~`K'$VYuvH4_<𕱸EϜқo>؏? =Y2 d(/~/V"ӂ5#\zcuNdT5 Bx/º -r-+E RiMbXpX,hP=lB4)xFP HDj5)|{13 aJDƖ1! mr b"ɐ")0R֢+R=157<̧^N#\uN웛E9r#!Q9Y6,PP9Lp%:V^$SsaX%rv鹠;Պ 3oGfeuiWOF]Ò] kZ@'&;R0`RIO 3BA@ x5"8}ODww¼Ofƒm 3wy]ф0!-;1utju}@s=̉V" Y@s%6n JU{QvcJxL[Oh*XPiE޻ief8 {H(;,-[g!)gcd} {@9l1ve'<7rԔ;uM5e $'Xٮ?g1/gcG.9aec5|).lA[Sި.6 x!}XPSXՑ#c;jhwqVј{i}&Orp*Ne!5X?b@0͹v4o$ ?F#r%$7Bج:S\I6%7eI-$vmYjaaZo8JG<Pо.Qx,(o"aTMɗ[L:'[a?rQZ@bmZBBDX+[V׈1k~KGqf3`:@ \w2}`m\9/2o/- re`azEƃ_ї p#azLj0a U@N9rʤ ȩrT;U@NSd՘b<@>Жڶ~ڴPv1 ̀vČP\D<bFr3tԠw}_ЎQX]or#@doDwkzIn4KfWyM&~r6G.wl[SF=;$rސ^3i6/3A$>'\S-s(҆ ͝ȫ/$+Lւs\4b́Ņ0&Urdb| EҶ~'ԈSNE'5֒c#Kvm.aeTo?6Mc$Np%"ɒB%:S['c!1pn4&]3M_67Rqɽ|i2D;/K`MKAVI|]6SWod2R7&YadI /;>xlF#O5&)1of"Y$b&RxddPG }K]a^ӟb>C2egMg"bM{ܣPUK_ē<;`C^wy%X^RMpKEi3zM%,7b|y'[J(P| OYyi9F;T l!9%ŋD X9+?|Xm`0AYvCkY,"5ENΐZW`(s;{/,SgvWk$<%'KB HB.2KFJ*xƇR|gr>eGHP|(3`MhègVvK=`?89q#Yj-1H7YNĠ$Yʉɋʉ6> ^AkB`.&wb$gëLF%d2XLc2~۶mڗ^f-{e`cΓ9AϟJ}vg$>؟ <6vg868~e=yON P\Ax8eϖ\ev|D7s+{gQ9*;}8yP9 2Ț@mZBBDX+[V+I5g@Fȃ?mH7)(ѕ$3 )학ՔSbK@n 7WOm,2#K$Ӑ~8SN!#:4Gb!dZ9&5cfM[uӟVcf3Q`=>*af0I}d8 )tV@ʾ! qe75SH+kA!a! /(m=qRHN^ސI`QCl9A> cc'b[!±␘XnB$-{I%i qz7yi]fA2%:;}9eH?ڽ]񠿼!M^/qk!;˂ZJP_'__ /pnqx'X1y7{N߽S lT?'\Qp*̇X=y| |7?~t(-P9_f;Z ᅒIb9U-g(/m7~. 6dX΂y{{w}\^YPqF<\U{f>,%Uh>nPP_4ssE!>x!&B; %: W1Y;pIBKMKT(|oy;I{l=Ӑ[,*ܻ1NBV6X+wvX&gNXKT̾1aRpKg[^v\\rj2E8Ƌ(?11Vv]/8uI>NR_P_-|٧|H,):MJ.G*OLw/}~n dXc5xsdlj+ Pȇn| 8$o@BT|ှѹ ziH|u8^>ʪ+Ep$ر։%7]2< \TtH= !T_۝fd<zGC\!udCo`iyxW2&5!R U Gq|2j1a5uax_A.5&8`||3 &$#i*Hb͢)aX]9)F֭ A.Mw߸}qg_!5|"}(s@qF,@bq" KdUh c/kɻl!PJAG]ј~Hql9UHV !pAЅV;F6u8œtS T,Sc2O|P `Ra<Қ=;8f:u}Ȕ{3ڜ+572gc;.Ft ]];'NU96N'c88Iu^1z +B-/K(x9@Ƒz 9F W+E&5ͱH$G\1zn5%y(̧ۀ"^yp$XׄVe5土:+B,JïP[pWP dc*e2 _Jr0!DlE:1s1{Id]\89 qpev%%H.P6Wj!|v`2 : :<0ڮ).g JÃ֬p6@;;$wʎ.  (HoYz,V[BT- up @mG;Tl>4?޻yJdR ׌pA1QRY_ Bx_zo0[4]Njtҗ4׽+]HKnL58fGa\+#'mN9'sh!KS@`3bh 5xC@LJQʿ_k_32dQ[ty&G+ - Q0qB=qCS[.1AC$R9%fPyOa(p+Gy~W/jjHir˖?KFBC$s2^y-T%F % !jVA*7꘎3z|g'9t-ir%Lpi;^L\>fj0&xvl "$G "(4ӺA P>.㢷)*6ۋ.oAxt.Y-?}$/F5| l춹$yN :O 1r!q #U0@ۨXJh`U) > cf%y`HaY{+ơҵQ|S̜9ѫ,*N/[ [G4^T5{vGeZ#^F4##i"#"" {=[7$a2c Ipe v<u00, :pXze ɘIpP%J8Ė i 99v""+)dlOҲ^zI)i BW wc =*hTՉu"(wI}^yX܄'78#(~F=?}LGCBR / U蠬ʭl-y7˯'w[4Ae" b+4pACz{u+#~qZ#6R;EwX9eF#$(?U;*2 P?x<1 wgB ?F_vTx=~n\$ap8=ŤeT\'EĸXg$-.2qfWe!,1NsE[F P'f&& rHBm@HrYM#٠EgD9 lɆYN-sAҝF?C# kjyfýph/4 d,Rc'y}MjBAmO<}Z@$Uji`( zN撊r}bܯ49L#ۗKw"[&un2zE;waY6"i",)>5%}/0D9'VO5 [R#OֵE}b4tDsIabɸOICN[ 亾OG>$<- >&nx Lxvɓ x?{n +ۛh WeuE6guRYr)+Iq|$%Qǐls#|~|it^A]I^~\XVڡ#TUxAHD*35~;Tyd}C;:9FE4qMcAVC_} S%B8Q =,t29eGCԮ 1e~m;_w{絟f`].>t TZ/VP|T8TNI4Nxv/=c,8~X^{wޑwo#]Ek8x{Gom\UY _oR/hoA7|W-e=.j>(\}ǖۋ_W*K޻>#?=>t5yΡ_.~Q{ۆj^ÿÓϯܽ9{A7Ϧ tS9/ubn*+vͫO_:$JyuqEANhc_4ELI$$Xѯ*?Y#&$Cɢ]uWb&)dRU1:uBW#a~ا63k \CX|Wl{]^^W=;{qo"svգG` bQsB4H8-D_F!}}:JHRg羢}3?#l^o=Ы l[[]InRsUv#D7*_L(aXT=59.g/̻,;0y H!q&OeJtiig쾢xpޝ~~mGNڈUZdC0pm\U]-:_h#1&f> JxjN MtemQ:ͳ~J;v!%mՉr:LƽnQm7gCtvWv"}3Gg|= g¼1 o?~l^Ykxg H1፴o}Eke2?vh,w`Wd{wץ~XCrUp/O@XYLSE*zuu?8.PEEM CST-)rD:;Q***tMBw;v/j~}:!o~{huKMǫwk<:^/VmSJY: IysZZY(]KNZՀ2{Ưȝl|zt~uPtN.~4r./L.C)~}(/ؤ\sU-awS*v0Btz9/I[p9xv2&4=됓ZLP$V$J&`$l1xYtsaI\ՙ iyr}*j (^z1<]_/ۿ_5fm?Kn׈?1LL+zubjbC]0v]6t_$d:g$oRE`d.Ob˃mI֯Hxޒ 2@B^i?Rqr2 l> 1PΊK^ Do8tށb$b![rbHGH^JGGL8ճ UM}4z.]w'5ـ ݕHfE5ҊzՕ4'@t xL#$tO|RVA,KCٻ=H^[aOe$Ҙs S֤C%oKe4 5Ch8-I ΚHK CPV@oF9sUS, #L6rGHOްN3"JѶb+?(I p1zw^Eہ#9YDNWyQ@#xRNqq P659՗8%!!zv2E1A[#-%pxc$tuogjaR 8 i[PCL@VjyOqOb#NCd}|XKv7tγt:/ί~{(W(-kV%l_T`dvJMFsR=~4,oϨ?]߉[K\cp%UKŶU%2.WU%S\ɖ iTbk/%8L5Һna؎_~& B"a!+cTZ5KϮ2[B_ۚJQp/!e_Uc]HDL%2M86\D(Jc-!#\MmqK[,M6l̓:dZ juRg6C9&bNj[AIL@{dr&K{JK;i {hX^ò/7.j0:-ϟf;|VvtduF{)%^Xk+HQu3Nh(izF" ?, xKM\t 9x`U I QHSF:'h)a'Xa @i~p+`k+M(L2CGHd[3T*Eڹ>zw!U~nQ`j3^Kr@H{ezc:VZդ-Lz{?(QAS;"_9Uuc$top^Z9-N啨@2"J# xO]SUJvN BD^ZWU:U;FB퓀4VATtPcL"c:^8FB$Ԅ}uKPbJEb6%kفJ~VAlʖ f5و4Ս ] ki;YB+czH 9c$toLDsZ ErZbl-6˜l~bXwu7@ֱ6Rsn*l~Rלs'nsqɢGJJy.GHwϨD}IdIHY*Y =;Nrj +equK ;eH7!U*` {vvju9NQy΂e"xgܨxaBou,=twr 'a!h") _ X**EuN*M."ũw6g/|:kFL9JՏ0twǫy9Z0}9[E/7#E*{",l(nfA;ZitŇ\|1+T沭#R( WOt |҇SrUVڻzk^s#]E+Tɣ^r.ê,5~ŷRm4v$j^}\|{{WׇryDǧobN1ϝ9O7slur6/=.6xdU4PZ%9IҰ9]S_UWW;*&t,Wzi#Ա %gsyLhM,(8P#Fh5D,=شZx_G}w>) %J@m7.ͧŸ_^Lַά]CNn:9H!g0C*&pM 1A)FR/v$&ˉ E\|W ë|wD FlůQtذݚ дXVe;\!`vzYwF.^Y:xax u2i K7^mlJMI&ms'8Mh>ɖm:.t=hAiCr !N=/gjqR'o: (Z@K ) <|oIrW:eQ6)ՙt5j f0&4Ϧ Y<헆9[gtS _Û⻿\ JG`3;(7>~|wvy cu ucg|UY'h[(T# jO3&҃3ʤ1hCE8`!à8T(oǢBG9EA8&6jFR[k)J,-|hUĚb%|{l4H8IpC`W}4`!THkK3$닁 HIl1+$P٧Gװ{Ӏf40 DU6e@`\Hc!tJ4!MA.,?xX:tMYb0B DލS$jHaфE RzqX* VD)@ĝ.R+#Q9=E*DڀaD 3Ѩ -|Lđy5 $DTNգV:UW*n,`<4aNﯪH)̨3>`fwTn[c#7&2b;.9dںAE='ZFG(Cf$5XaБ{+:]F_u-?Ip"u9c"I0USNΟFaiX{[Y)vU烗S{Z8X^aASPyg./o.駯D{ĆyS:fgvm~Ӫ s~ 4;D~9{7^Rd^V Ž 8{ozV++-%;j4wlH 5g.)v@^Yy֟oR_I%My 58VaBZOŎ:g`Nt2-Pj1 / 1YÜ( Z+&x0sڀh> p,$9;M|J8aMBMH$K$`L IMLS,1t4/w۽c csz$ "Mʂ'b$8(h_R`'쩕w$)ŧG`a]d@FJ̼"bXZ̍S"`x[+Jy}DR@Ro]>f@Y~(P+ "4ldoG,_y[xiOZ o濖(:v+-<[_c gk9eQU(Ϣ7(>]P'ˠ01~~ӿ J*.Oql+fj(˜ R/% қJYX3Մ;1Nb=7'A8Qhi,8*œn$ 3W>IQ2 "hixy=MJ^ΒI&$ܵtN$=-^@B;GB\"MhS2s,"'9x""T\)cK\mKPT^b>d\u#Md˃)xI Њ`%0Q78g1[$==\(0ͧʉ*+i t7䛳mȀՔY2;ϟ}5!%w0Ϫ3xz5>iv(y;;o?BJИa@Y>hLI"(?nn෥o<֡ǘdV0u~Ro: S*.:'2 ^Mv#*]=ۆQ99]a_<5roi@r؏S] bzFMH saU$Z7)EژΠDuS N#`-coAK: %(mΑRcXAS"'4xD  ]m)r y]_uMzگyOuϺ{w-L3>?5fXnuv5{o%4+kN]YL(gh@6DagxOṼvppu MOeeWcd#Pl#5E ~$L09N־ >^kƎH}{GUU9o/]jBHZ"^{-TH`c)qbD2K6mکNP 6֤O>lؓRq\Odo^9t@f 3|be4&I5Al[MDnd1at:`!@geOH’zSPֳ@B^H`kAu!= ۨmC–B,u2vB2f49]j",(NS$<(bQз\19 [$N)iG-a{-' *ի:՛ N ngv{ٟгW qϗII 5Br9G7F?xPq2Jp犅9?΅la!ǰYˇ4!s&bbIb.3&H e0"N._s|ǟv[%g\!T@͞.OȮ`Ia,uY ~ ]&dZtmG~٘_u 'kSֵ%n~8Iw<;tc3.aק67=ͣdt]o5b]Kͥ~[i;{M]6:?O<3r3? ~Kn xQfҜx7riבB;G7\lcC7|NwMOWz<)GQgJİ=&էvǛT=q)ŹgY΃F~O~n|q%#g= T2t1 yYסܽ^?8c"A9u,Hə*.Be2j$f耥ڄMku_D7z.jE뫈d֮ʡǒjOZ`ftY'7ܼxUEw( JaArxojT-=~F9XV?eztrrYN2WHf'``.x֝WN(^X(CL,aZMWf'mzIm6 NwO%kۆ2]Z|PtZBtKYIRZߔ-P|DS^y _ߒ3?Գ=t7]-bmf=pS]k(AJ`LhWĔ6y/ x2 fuDC_E_'i cnb@9Uv;ޖ'-F {;h_GxoHN;kd4Hc,qjk%N(^2 Bn JXDTk9@6ǎι= v׹mc%ݖ@nun;]EקY5WgWΉ}'p}N< `oc3ILˆ136gm#I`C$}-uId{TσRrDΐv ?aU]tDcrV0c ߺFw.:б^> [u&MKI`*Ⴟ lG=|`6S* aVjZ PsJLuJ%y48dz9]g1tfY$XLjAe޴qh[)nqvm y嫮ۿ(+OTp?&Y=1_v^qm:Ai޹hjӼ({u}ͫA`.GDhxߚh/9qP$Q3-9˜R$]3'Kc%y灬G!)UIWz΋iNѾ-p1|G?YKZI9kX_h1<33OcJk5:R=Taq:'_wYj@8CΔ>H6(/,6[3;yT|) AHRVa S끱[ D佖豉hj4BZ":ks/V73v:-q>n~MoAa`zt D((lp`Qi0v-7FhdbHjXQD|0arro'F },e.u0+p`ۻK+MMȯT%K}!oNl'S활ϔ[TrOSN]d>Z3I(#()8>jK ^Ȍ ^`A Ɯ4纐oE Ւ=<&К0Vٛ]|[/df<~IզPQ420)D%$@tY@^FЇґ0хT] w}CY1)f:$ @ kၱ>rAN=s" 4610BB2PPg($^2!YobA% * aK.h-;P@iIiMa=E P>BS΢ZU}6s\s{H;OQ5x̔JH ~%KD? _[ N&ي=%AN UÍp9&T[VPdeʸjO8#~6qNgf)$(`x2f|[gI99)5aY> 30!!I~rg\I8z$LXicyS ySoVKaspq״:XPtr HVHRZ`ҋvͰ1hpkMj*:X#5C$=(9^Tc&EX<ݛ)BpϗBbsVҁ(4ʘ1HVT&3X1{?^!fʣQ=*5㬑3ae%'b$Jf^ya1Gq);ͣyU-NM&NgJ1gR x ȝ-cxHe3,SY?;i0:{zbϤ1`|BL*5̫'/3FƇr7W}jmpUry؛i:Ɋ'dS[ufjWiaL>) L.C) }nya684~M^-NFw.HBKWd&w7k蕣VQ.ФBFpz,0OS(3L2FtÀHʐ߫onͰT֝nוW6?<`r0'(dVƐ!Y40/BGF{z ]]~x]~}D$>?J5nwwI<ӯr0}q?|vHUa,1lH X`T9F$㑅H W#RSFDD $EV 1lEWKYu9x ju= qwmb]ySe t*A>ub!饤)Q"8"Y,Xa;Ja8$?&w锟-ӻ4]CDZ0bH-_cQLbC=Ps e=Ph]]81QCMpr6GTdF]ٓ>{]B掍Y~R|{GwJ+J.3ˁ=TfhVi *E^r2g2M$[\jNRMX' NPܹJʛҢCġ?qҁo@o;VxZeUVz?J†&<$.(邂Tc3f1>e>|]/@ LZ?>܏sxѯx;DFg&QMdCqr9fhacϾ..fnA ZE34>@4ոڠGym.F)?k^Jէ=$dQpxǟVcFnū]I|7`e'`XQ)7z0~zۮQK0}v`/U;]4KV0#(6cFLaLuU'@S1- Ō93˭>S{bePG0&^쾔"AWڪ 惌+˜R;%w]ԫ =ЮRcg6(qs8 )&žcn*tAU^Y匌R qm8vHi*ǔ8 Oa "7h-gnUFH`HoQy*q UNj,O#/bg>hhWB(_y|UB!z SٷkL#*Jaʠ.2P:T<}?:U}\ 5F5|{%U*`1 `qF aI0 %F#XU[t(̷t F )Q13{vJ{ރ*;|uBQZ,71prT*QBJ;!NGi0AHQ%Q l t:;B!C*[mR ܇KGo1dxzWtIo,`! 枎d~]ES=jؠe$$[o}[dH*0wHsIM7wSwU|i(^),a>M6)PP+읥+<"Иܽ#罌T"/D R5SyGmjp-}aX^ǹY|H4o}o~̦1@5DRתW~EN[ ǾۇPj{_rT AxCՁA %7i،d3k,hD#ӃmKDmJ.JqTMsLlEw9lVJHH!cґ6[&5.gWY=39ضv}f}ZxJ?LE-i 5k+ON(s_a U9FQQ.Ujz 'Bٗ/#VV;n\Jv=oT ?6Lnݯkh|;q{͸h7lDy<_XܺIJ+Sg-=BSHkL1U9>D?!ŷ{Ə_c\R,iv )b-kFoB[B!0'm D}ȕm{l}4tm `NM)HV9p_&Kܣ[+gn3>Oz;=\2))Y#1qvB\(a`nh4t>&zΙ^;fHX&i& VWDbKoI ΋;{-mR~Sεvh*y|PiS{ї8nˉ`kiQ9 5R"!Hb)(=$锂dB:b="HI0sT-W=Cgk؏IQa|V?~'_'u\T󨘄"t*fG+0fhP6N!D,ە>U1_*v|bHn)=Q;j0Ғ`@ItqETzs#,{9$3yGVn=HVRb&y$lVi,ʂk zqyֱ8-c,K<E*r_,v wUUmqh+0}hUcY<]N&*"|R2 F1("%mR<J$`AhTxy AƏP+i:RyHitL*"a 3â`j%1@!$Pdb*{)=Wʫnw6b%&6#EPÕ8G B#rT*I0B;9VqQ rOuol?hjCHǖK[Bjصx/(Kwm$I2X!zܘŸ ̶aD^2I!/od/Q$DINeV2|"X|/՚@q.L jT4e6_^S EKǘntٟKʎb_CChvY̺q.qj=gg}Er\o] '\ʴ[?O7]>gW8}Y{jV,Rsny!^%*/avySO ~/`WƏl ܔm85o;od^~V5F {i-^8 <1QD!MK, h<q]Q2Mtiub]:eG0vRjU' M |hShw BìҘ;M|Ͽ'^B~pX3Ofn]ZOZUZOn/_)ږ\ffaY7Cj3g1Aw1ywep2 7iR3 ȞpjbZKA!l\[*U6=p{_~y_˟޼̼;7D[o8q1miuf4o{7͗Mc{4-ti״9 ^]>R/} >}}co_·!yȡ]sϺՐ5Za6.&sYw*dٚVKd2 0.سhQKdy'*)0 )%ԠAHg]3uQզͷ~i޷ɼH1ңE (9!7nC]2wwV/L5G>`cB:RWL%T3%LOPn\lͶO4be40DNFn:@-qEGC""n1@C Zs+Br$0H1x kEug7LB8YvUNqpnԗzWLpdIԨ|A w1=A%Fs \$ y 8OMslכaF'dY d- g/o}7RZ+nw]+X$QHqX X\c؄w]xܰmͣ;;@@IRU8 W*_p|[8 U8U8Wa;)l'vRN Ia;yl'N t',U̖T5HCI~;9{qF_l@} vum:! ?D X1\zmُ#= =gJ^zB/Se L)2^[!2W?qn}%$5V!ReJ"+ T.{9SVe aepc#xeR0O:` ,IDu8)R:^"\0H(]ԂOꠈ ʄX-hdžmrha/Ir. MVwv8 `0In̟=]SNFrMQYȲR)ֺeKF:\?1K $JB*2ﭤ#Xz, Dé&Jcw# (B<ň8 @Gbhf2*m˩",ٍ=QCW43;= ??c_ bݜa\bb*b.uzgrjJI5 ?.ca6&:DH~:~3L8(Pv" P2671ВDȔKb% yIB^$% yIB^$% S;opEgM}>uN䅴YZ0tdHU"U[9ld t4yRW ֣ΐ Cϥ R`i"%*ɭ5P.s=͔K߹+G:nf~^]zPLmw gz@ UߝU$*Al6&)Ю]{p3Bg 娆~> \b0ۊ\mm˭Q}sr8cL 1'(wշTHY"!X(z Ɂ`S1umٿ%a^Tpn^D܆X FIbAy Ap43T#I3-QNPַO),\ E>d8,i)D`(͸ Nc_}mugg&o(ևwkrrCne V':Ɔ ,Z~}iP&'czXrlGMپgHԪV~N!$H+cB┡Vx#uHֱ1U(T\]M>eocaK۝/g|1 ')PA7'E4'=`#9Y \iSGtBlOM,'G*צ b=Pv'’T3 ظR q1Hb!½6ܢ\H2h8v`[h՞R %𢚌6̡.wJh^TPY-Ea (O$v% ma C ]B}Q{d$ќq>|l,%T%eI%J#TMDPR T- g%D š;D T☣!"NBE0A9(Lzu`w9y\+^IPQ~}uJNUOwnѹ֮CUC@h{չf|?>e9 t=\jJo?Qvlf]!LjjMwW>%ۣ楒!S\ootŘ.-i;*14_&kKΆ.ќ Vtw6mYg e,Y<)HmP9UmF9U_<Ϲy~}dϞ?-,ˀB{xD i$1cet8}zH ?\#6:Ţ |H.o,b4.3'k=z]g%3 q\kELQ2J51:L9U1hexHhK%Y`$VNXEEM}#D(+| `>d2ٛ&$ u%$籇߯RKR˖:cwS|U"r$4TgA ]4| E 9R0qOBCڼ՚ 4҇h$āRQK2hӵ,6DÓ$\aђPZ#Li,!D-'Ifaapi_ Ä)X$$O@Q+줉 |H%Q+@l#';rr˲d9I9&-Y@>s޼!f0wb?}xk6؝ܣ;o~p\އ\Vw 3 g%AXPܳP,CY&nͅۮ^],;i)թ;-]=ujNvvL ݮ^S;%-Ϛn˞5Ptrkn!]/[;roőGZ3h<k4?QLh P --*2yd&4iyo-Uܙw5Vԉ.l}k}ΎϜo_xglrXQp<(pAl;ӆ_H+D)id 5`~¼`*@(I<g` lퟢ'f:~}`>g0 f3}Opz#_QE)/b'8>jK ^Ȍ ^`A ƜZ3v֝ݚq 3SRܭ.˓u1inؚlyMf ~&'PTȠ53 j-$i4* .2t.D'[ dl=! \H☁;ff29$ A`%c} Y Bn/eˀD7eI˒vLI{vQl'<2НmY9aa4j?Vxdǀα^"_jY8 0/"kH7ݹ?90 3{n-LvA*-kVP3j3jT͘S5cflȌ 1U3j|yT͘y cfL,Y03jTB1U3jT͘YeLՌ1U3jT͘ʘيʘS5cfLՌ1U3jT͘O3՞L'gHADkr^yA=bqfIb%F9s`Ԗcc|کHZj^Gޫ/$rΚeeRA3hڢRa ^ JIk)e ņSu(jrlb #52(B; E N&x;6F;N $ь}#NU^JH;`&avNCH~&9MU)Rȴ{ G1T Tt0`FHA^2c],wխ a <" b OYc*h-µzqyֱt֝MSKvkYp^hK-)pxYM06eB,T(=6Xőp.uXE0L15)6QMKc5)@˽brmjtJ2&*4>ԝ ߊɜJ@ٟUŋ|Q 'tיkU 440j1q֊=4*aCqcLŬx 3ŸUٛ_/V ~/;se߻і%mWEj7 Iե,tih [jj46ilfYނ 0V: (V&df728FicmodSMchurC-#aإ'b" Y2-=*UzKx`p%~_~}~}o?b>ǷU4uj?_i4]ꦩbm]zN6.mVm-J(PrMf{j'm:-WM{Jz _ b~Uyiͫz׵^ 0vrWR! l wCt#{8Mn%sX`f4)z6F>j6?6m%`?+à z#(wR\>~=ܩ8hHUT3W$<"6УHrYܜMǽJZRqz毩"*`L(xUq_@V 'B*p񣿞@ 6y~VT fˠmʹk<{Lsl+]ŋV ^NW^ x{^~MXŨ]^EoPL/C0D|Vݦ,"i.i+[,ڕ_Lcn\r"R%c6(SU !}v`u2=T—42ΈCJS=āsñaȍZλ[cRV fH  )Oc.*5XfuÚBz P-r"k{ZWm/qم?-K1{~1:ꍦ}&2ŀY/;09-J-A"!t60('{6GֽIP0VSDq(cE+N F0wjQ1G EӉ#"( -J)֜MƮoK;Mu-tB&tF㐶|~y_&++6/y ԯG3}|\ `/bC PaZ~vIQI)`-etDc2[@q |v##->!Lݏ1@-6uc 761'eXo}_x0rf1L~j(n~\iAv`eа-!lw?blt_ m'g?>֕ ZPkekfKa,-Mov% .z ʌy LqˀY,C2NcʠՊT9T1^0(PPT2㺼J@֪&Sry+EɑjB1m46n nkeRoAď5KlVk옮rRv]nB~{^rD-P.QK-9+Rӑֲb7GƘct<ǕkZmAj٫d~V 3U;]t<AjЦz}vl#5˴i]ۤwpS \ǚ}e02d=hhҸQNm| tW`aͦށvm7/ Cnx\ > ;lSEٟH0neRNLhS?WVʳa~)sxgyvA:bUgwi|Jc$J).Ģ F$ixGR㎄}ImZڴT urrfǤmٻgȍ"䗶@PIz9P4We%G8wWOe{%9pvKCrp8M u4mFmڜy eH5Wk+tSe7tseyP ?b,ߋbX-Λ/S5zyREC2_\*{R=|DWFn46K#HZ"3ITܗ/2I.3ek}ۀ_SŢBA[_LF9_~rE0sElp:R.=LAJ>e>gD9yJ 2BclaknZ1TxPJpڂ285.DH9:5L%ʬb X#5(O @"a1Ёׯ>kE&nng7RjoPuum{K}oSX#D8FZCyx) -ZV6X 4=(M)-.ZJ: 0J "uH%3)B&I8P%OM]Wg Uu[3mܠW.GrT;ڸfrBr^?ڀȘG]E! !S9#[h N5Q 5 ĥPލ4#,aDɨCW2+tB?1 dFE½zOd]B@LN! E0)+7o]3&)8*!"/gP(JE+C >8}Ǎӗr@J r&T@jU:A]tDMj! 4[aPAkPP`1x i93V@_C\jw4 n%<.3<4vOL\B@u5xk" h"'H3P[mDxkL7j/d] ej.AXo[8 ,)fW|爞Wrw}X Ş?b5??뼞oijIy&0u 3Np}t9ۓlucZV\qW%nEeAg_ÿO:j˅":'qg;wLuotF]B*>6PݨRs8`:Oݫ+Tk5i+ó?Ɗ7)7WQݼ,ct8g8ԩMv͞Zΰ}ί'^ˉ|$tBr;NI-䴵&&0W_޼#XVwtm >υOh{i2NQ3ɒv;ɋnɯ˒V\:m2qsO$H]Ii&4)P`--OĦ6&1/:s9 vƤe>'ΦZ8a&\&ꃳXRy/c@ Vc*8sj ě c_ףOB@wQ}{Q/~ylWKOo8 ?xcǏ{@ɬ9BX׵㣏v '<Kߐ tpQU +hpKCWk U>7T$eI!BJ)$D2C Z:dHAX4yHa <5@s4k>Dd7C&(GIPu':FQ 1PcprPv )K$)PH,xJ;>nVkLoK5eCH=GW#vh2kkNPԺh/KNw3CzRu ^j\։vHw;L]qm:ɌN ߧ'.KUpG[*~ͣYV% ,Yܺ{Nͳ+;E]pQ5Y#=p9 j-ZVu.}g;AmhzyRn3,-caL;_Y{,o0ž7{ ZM+7ORnvsr7V@sғgf8_`(VVnn]OQf >Z+B2[lxYFN,XPX(fV&ƴwݽvV{mMNx}w"mXLׅV/jv{Jܒ ]\bn!Xn, 5h4orŞI#-2 sO' 1_F&.gj﵉՚mieX&1W{*6=؃&UܛwJsu& tҢ`V_g<3;CQaې wJsHin(7x`:zP~&BQp10` jq|}DkUw ɪk[]8]^w\oeiQNh㠂 Dhp5Ea6ME929{LN'sp6 ?[&'Fh@чu]ϛ_RE=$‰6ʣMKLh:PGMf(A(A`oLUR 0J PC*I2ALz*y Kpcl8qVx}I'yNQN^W4K-Tk?:-'WAG,u<"( a)r)RIGb=c=>Q@4jj<K!i@5FY:ÈF5Q%ho=9XXQ3~|c_[yl]a-<4/ukHnR8VH~:I #I} g>mpFSgJ'p Zk ǜɃ"WLpdIԨ c0sP as FK7)NA@=mVI (Ds-rmJR"t%Y0M*"g;^Qvs cFJu)zi(DɽΘY-&6˯UbZ$c":$pC2k^x%@i H.FYY! u9iH# &3 l *D&F@肶4KhXV#gǓ)I %k`EAYԠVs Hd8˨BnbC[c$GRI)1g1KLG@b Y$gVzZT>XJ(F}FOIƔV{'#J+V* )+ nT)R7d 0 !2vHFCwCȖAV6n0#Q=l?@f 8ؐ3! !U*lN F]JjjзWyLe*/J~tq.S~CAq"B>K *)%16w2*_y>pO+ZNs)0wk5ӲP#\{?5$ NQ%KRTΒd6 P1ΜE*@1e}&gZE `P"+Fg51(Qq%@ Z@}屲 ygUVٱu!090,ϵь〉Z"H6l(̠7z(E︑U`;c '&@2J|:qMB#Ԭ =$jzcs(IY/vn ~Ņ"SЎBeja}@m.{fyw_raLqq%960S.at>,`FT?/\&oJSB %y!OWD|,)5­#NAe :儝\s8ܟh/]d4E$lQ5DY[GDb*PhWk4nq@ccg)UB Rw)v\HQ‘}94 t|y@L2.l2yw*MN_5%(fڵa|{n!qqJCgOuNU<cig]~z s __Tcll Ej߮eЏПn/_"tYږ\ffakg(!#G1~v2УYxCk[edsNku]_fUTԲQm Kd>W}=+uEΠr~JPBWUY_Ƿ߼{߿㫷?Rf>߽]O8shu$Xh83wo;7Mc{4-u=໴kv(00#W.pv%8j/(ۂۙڥ3Ozd:CKc\F)du艇ذEFı o;I ^{4#7{oD:;oH"!.,bxuz6J娴Nӹ9`ir;B]3tUGչЇА!p 51dR ~\ 2Ci}/2boPoOPO ^۟td40gi廝Ye]Yu1!jkЮկFP[5!Dž poٯIz}:Pvq}p=x":mm&3[D@4: \p# Tt\)"͜;K:y2_vkY=\KF&eNDa9qLm.d˜ U,#0"Y `|zBY/J 7$GI][BGT sSDk|BQ"x-wRUƺ>Ȭj|ȬeU#2Y9,Z~|"ē'^?emLņO7ݪ]m:f sz?k!Rg`9˘e lN)#8TF) `r9+ 78 2X[3V#gJkQ~Tkq3uuudMTPTW/μsSw.}|:߆gw }ڂN$p.hfK43D8Q,ud?o? ky@_GR=Eʌ)Oh>=G2eJW$TbCa2bLHOa+%ЕXjq ^YK QXiBer9cVjt!'T<` sO3vVn6v֘⯿|*hx$*Rh^q.E4J8fY(Nc0YE=S^4b5xF5hiVfh$fhZiH@2ͤɶTmď{}/J*TM!'a78I]L߯gb&Y,gr) FZm!6C"D,P"d.*a .$LK%LRI.1W%4`^,OKb4'=vS, kGSC%]ABnóbQ{Cd4&m'%*e:g7,|b;?p6œy.TZڬ 5IC_E_˥jjOo51Z3',V@'el㬨*sosn(8:#* @B#ݍ9%QkoV~qb_mW4r?g,d@HcH]M~`= yhPu1B3rYzlZIn1dx`!ҟi1D gXdN0Kf*j MNRk.u.-ϊQJG";爐,X&sjlƞ;؏p>v׆_O|Il=?x> ]bޫ2x#r+O@iDǑ w` F*)pk/v 8?lpg8{--}>Նɳ4B$ JbeLsL&3MFr2*Y gYIDɀ<cGZ2Z$ ^WYx17tY^_s_WK %g4D]/oG .nƯouYy'琡M14.GK˳h <KVA r m-mȅ`I*KOM|THh5B{K0$yN%UYVCrH~pI:(x1;k$$ 4Jg)X&I!IġBPWX@["w"NRezȜp6d/0A։[^oâr5cnz=m7:$3{AFi5|)xhIވql$X qli226X R6Y]~lCIUh8;pO@Lg=YҙE&wg.c|jο*^H>wfWp!yBs*D$AdcƖA'5fEd9v9vnx,wf+|flUIM,N"kS( )"x2QKcBo,Mjd h˹W:zgE2 =U~ Pt6*-۰?ϒeOI(AHYTEiIƒւIZyg봉^ܠwFgC"S iMF&11RŔl%C$^iTW 3S(Bhn !s?KᏅͲ0]-Šɚ+х#rRFb'HdS}`yTazP|0?_G].+3S&K3~}@1$S5}#c8-:L_Sk2txzIJ1}_[t!(}ȼ!*2qrekZ,,ttG"]97߬ocH0ᷙ0S$*~\7D>\~8Y~#Qٻ޸n$W23X_*> $^#ՈJL=+W~X:-#ҭ>&WU InݵH!tCz!B,>ޖ~;7/<&gW N,lƶ[ Ep}яxmre%h%[:jF 6C+vW 6+L>;>|2٦t~MHVU'Z+[eĺճvHk>g_uVĿd8Nk?5?j?6=^=Iӓ_~O=m}hlWR>_`yIO;Cq7{í[choմvty^/ ~vݘ]h郖vֹ@Z6ػk] ,rpNs^mp4|?| S\Yp2 [#HBJT5HYK)b":]?WH>R|+HBB1OXc;Zk(eHYE 5&guIh>qKRԻe+qkyOsފU^B.yݧ|Hm%uDX\b[p\S.VtVعg>b*j|.ǚbRI8mBe RcddID's3O .n ceu#rje~:Mg7oiQǍU/T>19~eGW|xv4=79!FL|0%ʖ%o:(a#JxyKv;gFk8} ,(^'bAW'gLC>ү[W'nyfd{[N^GD=/&z|r:/>hwRuҷh#~X=i6߾so}>.罘Mq^qMߵ~0yQGɧq$HgòZsj"ݛysoyԱ[ηSgY:sZ@Da}~=/j @9pű6!ĤnA?9ezr-%*{o0$>>zj,vNsHdyradlEieh6'Me& k'_5OZq]T;Jwbhm:cm]mid_GÎP]ϖe󵨎ȝL]rN{Q' běL$>y6 r7K?_̠S~y {v~Zv8:{Z{rtһ-tu6wbp&(RrO0oڕ'?::mF6{\/40grl뿘9 G ٺ$owЫ)K 8U-V_.uㅾ^h˘ C,?].?zq{ 0,WgtZ#ώ׬ [gO5=ry$7w7u|sO:!6Oйp؝g[]޴ML[mNE)Ĭ1lptvuuV^;?{uΰ~ۅ)}>†ƭ7hإkg=Yf~+FꪃˀqEVCAlP1/#[020?OA@׼\lm?4Ly`Rv쟫_{їU֊]-8ˬ4YT W4IM۷Sf{ .](ӷByf?6ǹKdjjMVDXl6Mdk ;zW1Dh}&!i-BŊm1jW9,!ݵ_[!cKVd)6\CZڤɖ*2BKX FTrB j-y'rdDkK C.:K%s`xITB7Ue1x`̄ J C@|!P v3WŜE1N0 sG8_@A.-Mɷ%;cr U.b,V9"3i_NA 7e SaBQ"!\`#)# +Xղ$,"=@W/:/()ER4Dbk ]5u$ElwvxFd0^2B/v 4H ՘`C- )$QV +\9$Vc+:c"& Lg,`UhVKo+f<k1 D%ŊԨB s@+߹OU[J{q ۪dpSt)\]h)7UASc2$3Hv9e E cZ*$bDN˄̫f䢢R0-ad1FƁS ^\@$6X/dځ ě@q+2ie ^,TɂNv/(ɩ᫘w, )8N$NU&vuhC$*a7]_GW1#TrEhe`׀{(%GAn!@v)Ѹc[ 8 T%N6G!H<*g !kFW85C0(Ʈ0Xd5HpsX(FaEYE E]p 1ˉ䳦`k6֜Or XVHG,Y ¶bZI(!mJ)g[q/z XPKOJh/њvZ,AD\2Lk1ژ[Q5ucEqOҹQUʄq819u;;0sav kԬVF0*mϤXHLT@WMI~B1X9lY%kP檜Wm57}Yw԰*>@f(m@Xx~ 0A#,ΐz΀|QbePm_H0""A rT`5M{l1 JIJC#I V/  .`-;fT 9Ĥb~ RLȎQXY 6ٞLID a8tAb鱓 Sl$di 4ຆ@Qv""DoB_3@]j,n[uC8aӝRt융SȨ 6B~}Gȥk E;G`dH g]?ʳܱ,g-~󳨿@B2c'ػ/I)'VI @'.=ҟYút~At>Tjo|gP(7[xBdǃlΜF) X%o(F}~pc0$ cvytqQX"l( J3nd.Ⴙ (|' ;]{G| ,~[V6nF6UkG'q'_rlGO O5f O?sםϝHKQH"M պ2E ;.V2%4\dSEdf3rvk^ʓ8@]hԅ'Յs&)fu:ۻ*7V4RAskl[brDaN # EFqؠSl57Jg:La0ks'dPǤ}aG9P̵c95~isy6lWq.ZFvi]Cpn]\$gH犈` тUG{ѭ,JP;d )O`,k| NGQ>R*t;#g>7Oa<hP:ֈ9jDԈ[wn<ٖ {gSH ujr . 4cpNj"-zߩFڡF쌜q0 9uv%Eޱ^dQ/:\y<<澚Z<_z<[U2Zs^AtK:Q^9 3VtItQ)tITk."'0p<|?Y3r!)[E[³f;N־YEseB)tJ_GIBe| t0E*,rD5> tryxvSrwO,$/Y?ށtncND3K t*f6,!koN:dNa:;`ïz3S3޿eDn%anLrwz w1xO4d:Si{Xβn_ͺChAYTn[{=/|zGK-x4⒋9{{Ge>BIͥY.֜Mt{狝nM1NG{|07Mhcgvi[poDuMao^(kuM)5źXb]SkuM)5źXb]SkuM)5źXb]SkuM)5źXb]SkuM)5źXb]SkuM)5źX5ߏLQII|Rl%MU~wPl(->Pf +cX,6}aFλA=ZVH<{e$N^ոa6]]%{Qbv%%n<^iUL@ox,^['0\-a塻f|cP^{_)}jʋk?d7.v.rY($smg"WO,X(,!{jZXz_ՅItQ2]@hDvdmqIwj> jsrkG64P`sQ5[ )&OLҴMrT')MT05W[(ֻ{fe6nƃOSl3 OL+,дnZ wcը6QUE2Cse9Xw`UZ2։.+7qpu4*q8vtp &%^PMP-ߙ%^}%e $teA"LQr$ɸ&&/#`4?އ"L2P+} [sULno6%;\MzPY%1h4%8V_O uO}-8|gQ&sn},DmI*0sRXa)d橣:]팜,h8pw/2ۮ+=| <ȗ[:N KFcZggg7ꛇHFh0gs^*rxBea+ ɔh/.VVHݡtFΞ]s%E;/Q<%CFV4 B D<,'V .vI$*@ErU3FVݕ9ɘTT0ɔvDCy)92(-B;#{ Y;uMTK+x?J[8s)9qf7gBٮNj8CyP.TgI'g0/tRfVxEȝ3b:ǃv|;Z^Jö>E*jxĘSYEMNҦ2IVÆJM#Ӗr9r=Tsq eh̝f5ja݇idJ{;LΤ&T F97*,\p [%@!? 1hAFun_^3[G.'S5y?#ȋM16+U5"۰HS2_ܐT2wa*\E5%x}:>$ T\[j!9‘B䪂,w ,>$. $O|?_gze\]~ wS+*~;m_^\\.\딏 ijS"|j-؊>0pjʜ{Mtu4O~Yv6t/6_e2Xۥ"`ۤ˫7}Swoui u$.6t6L: 'y0чrT(XɒlŘd97&6j\EK2Ja&#M.u}[t"~I/ho#cSt*'~m:o?|O?߾o?\}x7Ο`&ah#AI8Y{Cߛ-Cx -tWu9q<ܶO_^AOtZXͦxXHA=? b~deT=7kTIpvTk_>a@۹6A غGZ&&G }Ne{Swq~}5v1m2"'-r廇͊}c:vvi;z"HpҜ;LTºĠwBrX 5^ N,1H.ά{ȑ`fDق. D\i,.N;x+V|#D. 6(D:r*#9qRr Y(9]u-tF.;0uJ['>Mx O?}"/hD\'s}\φKPWa}PJ^H2Ƙ|!m|RbF9/:PH\>`FFYwL]zOv6ߕ'ʊMOs}p :ކwmm~ 2bwq vvb"9$~-ɺ٦,qnf>~U,Ve iЛ`L~!fg}i;@^:4~*gl{\pj~&ۮϝ—~wkno޶Pvpցw2}vupkZ]\hVNMTi2n'N)(C}Cɲk2BE%K#r(;Gi,zkK8r[r=rɽZKaF,S\.K:3;ɓ"1:)TUgđw\߾I\7jvls-n(Hrf?5{6ipTPcX&dI q* m.mx|7,f!x>tZOKͧ?6i흟7W^nﺹ܌)~o7fpW]NW% }𾇳}x3z4*IZr,zW)m2D-!V^%߿ rJa/mj$*AIIp+Uړq@6hTP ps.dIٓiNsYꔷ+OӜz7ЁKY` +mQtn0ʨw .O.?398NuF>#3zsd3xoD`Z0Td)\ +R6%*j~e BE;S\TT1\K%ӹ1&_35bnS $ _»g_Brx; s:1 'm/hiD3Hi"D62;j2բ!* Ag 8F$"pGTx`u6+ Zm7ˆZy€>Gp) ~Unwa?ι5Z( VyXA* 񓾨pMeiILf"87v^U>SySFPW|:(Z#cYsf,BVeZrP_eq{ W\&-|f?uy36Ek,cIIVaXά9% tvdiǧ)\so5g*Zɤn V\F|`wbxb-Qs<oԪpKI-2H)zɄ}LcȑG~;e]m8Z@ Q{5l;hł<xȆZI(A+4)8FV+Bpi7(k28U4]rÛRH_ ܨ䳹~n!"dKW9u+sZ9d9J .%qNN;yxlmljɃ:ۓɠle:ja43\\dJ| DȢn&DmM1{O^6f8BO 6XebϣK,IQ3-cJFjlG, XXmf< vXW,TҖ2mԞ! SASٸQB^E{axY_g2v޽z{!76OiR#׮]^?{\8 [l<iM7w9ofloͦClIlStn7 o=+Cs-7χpZy9S4zo8y8X;5sO] mSs;DT%"§ΑvM%7ɉ4[g6ǴmK>Z@,X홞M@RfHQh4ʯ.IxqQѭw\߾7jb**M.coˏif(v^Dqm LY0crYDBr0͓"1:)TKKN:l5]6dn3-t)͇vҒ)4nMO%JZf:(!:(M4), #BU\+42̠ S g' ޔSe|X)ov<,\(^֤L.-[KTYlxܒqǣ 4sfv)sɲ4t^}8buJJqt\^\L]WNYJ %6mm.dնn=o1-;{Z\rLbȲHnpD/nq^Yv,غrl؞t6GW^3ۆ{^\bv #ڨtey;ͺxܩK"D vұvfkhZsQ@yU,j8Oy+9N*F߀^w CRa5&EƩWK\,d0 ?P`|oM^ ŤFJ!aJRfZ"%U'RIwI)4O ~G Y*9R5P 0'c!{n}e98;,k8Éơ7; Y*^ֆ΢Q //F\mt7];vtvIώ'e㸤@QNЄ[+t<^Ɵ^(DϠDnJE('OMΗLP'U`j-Qb .=*alyB[J)ƅP.p)nWdtYK<km@6 F*rmV5qSe:"JOkiヷ g zρRN#\ ZȔ!<2 Bb&egc@#ʉtM;],]Y7O.)r/ĒR/:h]D-i}MKg`4 tHk 8EJ )#V@ *m7Z.UT}&RJjUT !@R9st4S(-bU}q{@s`qN4lLlt \كv ;pW$*_ܱ T PVTAPL1 >vή9e=Z Ly9%w"D)X2 7Q1s&VhU'UYrZ='-כࣳ"JH:ļ>5y՘MO^*x~*oʆkV2ش+Z1 ܶz$p Z[ Ζ&‹BsO2Hl,c1Y-י`bBHF`^"kwkK^2`)yO' ұh-(i:XOVȹ/٤? |uqsPҏ4T#q\vs^^ۻXa-^gctN!K"I6GCAkY!b@ ޶++ɖS+chYD9$! 5)uQRޒ* raHƱ$ٲFΚ nBm%ŕ6DII;%;;HC2 aQj,8YPR%)6KJ ]P[{s tD>2f0`)!lG@k XnM>;g;h|\ՌQM·[IƴeD/A0hRZ227b|5Oowkd)elDs;b-C[ ԹGr~qϽ_[rE.UHJʕjOkֹY:_+=6}{]8L>-9/2c JI1uTV"c ,alTLOGbOk;N{wi՗+-3Z|gl{i2 N<+R( I8<82"s Bn,th(4d ,^f뤊Κb&"COZ"FjނW?p-5Vs :ϲW֥c J(5IXnWQ'%һJGձAVK@"R *C&@3), $M# ^]AzHᕸ6\yn 7 .y\h/ gAj\-d͕Ӆ#rvf#iޟ|<.0LN}Ŀ̤M˼=eTy= ?5)W'.d Ф3|v'< ɔ.\5+o5Y9|&*x;&g pZgL\o#0|oԞc (Us4{W4k7#Z~oJ^?M? n! [.XmlZ(z;A%=-I{:Y _ڍ1pT(2e% \?OW N W{rYƪ װuR6\FJNJq,MMH^]5oҨu+46Ž?DW|~_o_?2n_?Ӫ') l NnL^]kq7w-\kgN!/Pϋ~kkk-@B|=0ҟ4{|OOtX5*U0+D6Is?Y5UF&sAU}-¬@* VAf d]M6,XU"Mn9Yu 9<8UArQ:G-:*ʳ ̕ઠ2W/Y.2HJ[nLq'։=%z;Z*h2g%7Pr|&ZfRĹ:WK|"QfE596e86ZeRv',1%}&֌F(ynԅ tQuႢ2Y /o kw'8LO~SA25T-C0}*Db7|2œz\+o?k##5Nu!Wԏw+&n?dq; 5Jb) U p:%*䬤T9d^+$ӑŗ nugO42Ad8]Kl)ʘNNfrl.lxe[y]1^Br) sA*+e2.pIwq*obgf`ɭ]켣'Yӣx2oFNr4l+A<f2ΡfT^6pJrQSAeZ"Z 9 Tyl7`j05$e~TQm:z-,1Tf|Z7̴Ÿ{G54?+y_ۮ/j^kL[l<4w:z8k2սSc7I/H-Q%l-B?`YavPIsJytY$ ;Ǽ,}omE{.Zl-|L6^ ހ.J}dCG Ȏ4L6hTQȣ]'{frum7Dv{nXegM/sAJ&xu–muӥc;%2_$BCE+]ԮګNhtQgFҠVf լ ͮ-_T񙱲F˗jhy!c^:EJi"FPO5eNlOo(+ e2Cmך KgQQ l𩮼 S3nx2=id==ylGE` QSΒC ct٠r :QŸT)(xh{ IqÄK͂ЁqX@xD+0# +aUN kli`ʂ ZFMT{o ۢEͶ?0 ov3uO{ ;`≀Tqg[K')A3OiCajMxRN»=/Hw)/$>\~cwel]`)v?5lb Tb цowp5|HmTPg:aZELqUߴD*: Cer=7N8BmŦe&D7F\aj4- {2UJsH|`82c/'v"'N̯I_#^'xgDZ! lO9n4h28, >HWhyt=% t< 7O[rC3u7z|.[Mx21tAΕeNF]2T X̵Q*g er+gI1v)Ժ2%;c#T42^`qHi*ǔ8 Oaȍ.Z˙ێ~ഈ[bHHžYsT}!Ju>e|;6Ϝb˱H:RL"hp5K3%Lr6D"/aN(Esfͭy4H{hh^EʼnE$j\Y"CԖJdw8 ao`w1K@JDx[t3Kߙecgh7jY q8Y˶F=4455-=Hw nTqDv98WT\!.R0Y!D3aw ʽvs͐L 53MA("R\r{c(1H|F_\CzvQjkw߽`9?b}DIŵ4瀄k) qZzlk$Dl PtJNcA3Z`(3*h`b-(z%Q1NPDr8)x8uD 7-}_S^| ͒`-,hzZp6)wɧ/hhliGL![amc3d, kxڣ2AQ,hl]qeB:@c9"HI0s#9:ЏDLךiz޷O?A.}w^| "d|MkZ>ADZ ǔ :׈Հ wՖ33Krh˖p9 ju$ƇK"}hV KnKa!ϬَzKaJ-X(ƠFDj1jjpCQ{0rHRyHMLcSB )&8Ip׮٬޺4_0 cEid`Px')]kSi>ݪR*2y^G1T TtTE`FHA^2c]!X$\.%\׆N Ln8KvbW@eAev_(?!ZaE2`/o0LMdKs-l==ɫQM Xipv([al+ ϳ N-s`]/߾0y2?'ҭmZt"#$e}p|$1# ^:݁i!~\j"%j׎0e7g\D85(a@."y:6ߜ]E*WM|7QǓp,(} sqEA5ԋ()e9+$q;O MzuwF9K4˭5\fQa251(hoPb& T ";n겭KG*#R"%1,Js* J3pRAy"w}+=6nWèRGX۠?EC W{5 CQ$p Rq0 ϙp {?<y.CHa,U\9DPh(H#罌T"/+=!uKH۴ o"ofY?llϿF\}Be`KS =hs)ѣ?a!IN}+5R"qYL?\OnlKxo'\E}L&~Xe+Pc:9RL݉Nå~=΀ ^Ϋ@0! H P*WIt 2p"Ѭc~;>3(|.ٕ)IK?ֿ-?;;]^ݪS0 e:U8qނ)W-MpIx w\! a}vkMYzr򓟋^\./7\/Q/ʽfo Ej_gEf4aKLA31z˦iH4 i-q`.|<~7zz3r0[8Fi^dӬMk咔ŬӔfjIs@'. }= Μ}g2K*fP8l0*?yW?N}o~Ëo?`>凷?[;؁Ypa oܙo/m=5]ꩩbm]6.0] ւ(}zȿL˸]N[scF>Wl ]9fKP vnr`! d&pn}ՑnKV,P,tYfKT>1(N1hbuTE-~hr}% ?+à z#(wR^2_#' x^cd $X5%2(b@4pe1B02EQR+ums^h!9,ot]CDfVkG:N03pz0dIJ Չv[Ũ۾&}zѥ7ޖX%%^u$0BM%R z%%Q}%'X 0UYIC V+C5Ɔ(wXݿ SiMyup>V͡RHmv9%uJc\ ysK0N)[H5j+3I[a(UV^Osnۛ,(IN.i_]|6Ct6KL|/5&c9k wWRVw}^W;ZsY%qHzd*Y(ٱ͗Zr4W-`'ٍggAA:azг]xRARJW|W~bV\5}"jU{]Rh𪫘Pa[o!z۽[tn3Sq  [RD?s9ڋR[i/ :UdݳmCňЕ,S]~n5~?|w*oKq˽){kZoQ,[s|,>$pK7|!$k)0׳6q;1UHKfv쎥ג-X#f{Lfe59% zR矪b XJ}j,IO3HS16'kol|OוP(&@: }LB_ʛy,k-E dD#eΑ%mf^,eUd4f#+odA&}\*XÍ k2X*2#PX(Yϊwof| vY/4a4| g/,AyV Y^d" 'lRɬh@046"Ƀd1( 3KUFjG8Úy2l-8=j.Fz)lgABtQF. VcPWCu>Ewlkv;UG~'\Zuy6E{ KpWXbgtYV ,&Pk]**A_QХƫg;"a˓nTͯE-x z fo0^|!xtqoZ&"2F-&[W##O)zHwo_؉@p2r|YVW\ ǝ ׺dX`2p ,=\* m[y/դޱ }>T-:dPP}D=J)m ~WW|'Tċ/}=X;a<SINiTyN!$E Arb&DDQjN0^ 7HXOI;f| w߅f<܈T{ ;fFx$,7D{$>l]3tᇥm ֎:&7:6pݿقM0 V^uhީ #W__.WnaWrz'^=1q ΢>JdsUT\tkpJA P^O'N\S~y皕{(9ny$K.Hh^'sνϵUlw|,R2>5‹TMeR <Ov'WPInV)S뿳ؿTUi }d;ɈdlR*FXF%T8><i<"`>*)DC0t@ZJlnfd4Ly2ޚGP2L4SH_=>,!v5(>ꊵ.+5rp}lA:dI , <V0"0Ԏ%K"r!He)OF V[R #X3xK4T#OUSKO}"Q b.*rJZ  kUU꤂Eđ =bSXxP[}+jRVv}j,Jn"p䢯1Ў+UG}@U.:bI$ n4i}[fC^8KE>L}?hx`ѫq'$7ߟ~0B.ya( B\3X92( K_ukKpBe!C}*.!6҆[ &7óU+,,w՗-)Os/xz97lԮޏO۟3^QeU8|O{13 ^z_~W ~WDR"^>?__ܢqރ{MPǓ}?y457WCn0e*̲^&L2J T7m|-&zGv [aq+'}b}oa񾛕ӓ6tV\uZu~>\'2[ (pKRAy6tr΂뺡/p09OZc_=vյ <˃YU1L#zᆰ_b6hv4r#)+nURfy2?Pӓn@.AgvB\(a`wJCiJFR<ҭ4E +xSD9VpkdHX&i& V=&F,B0#BXYLLq1тFQ4`pHn18[zSOq(_߉307e3s7.ly!˰#H*AG8I FKILF왐%i#(cF1Z 4wa-(zLDyZopo Rzx2b]Ykׇj"W,fy$RQNu㚥N/IQ<`8ڀHƖyXBfj2N{N{n*qV,qTQozAO,|ֲ۔4)δ=HX ȃFă9Leʞ1CxYN5yi_-d#Yb|Vȷ~Ӥ~#o\q`vv[ZJiEvdc9!-+j ]\p[ =uˎ> ]-+1i ]%ZNJ:tES6U,ek*լ-thNW e5CWL"FX 7צ*e-tJut(]}@Bኵr>P'joYyeKh!?VX["◜^EO_|MF X_2v9VH "t[#U&ou6HӠ!&2V%56;@+ %#ҕJ"J&0ZCW :DWwH(IJa+6q ,鿘,&+x3=(tg#ϊ>c;?RI;GKpykuJ( 1"b)n ]%\Z}F7@Y=&Z#vqC. #α TEln0E4 n1%9IhK `M@Vݬ?yޛw*"#K`RGQJmƌ6>f.j#H"*(nF\Mۢ5'$sB\6`T;=^ .i ]%oI(Yw!)+K5(sc3i sg)O"2L/|~;QXCգYCdZ10@9{jUwjK [C߾\e-Zp֫|f=G(eNi{wPck!!h$%ig_ҢuX8]+; sJv엂CC: |pO^FLtEabS>,P7=GM5K5ppCuIj} mbijΛ?/'M&ee25fO59Bc"1^|(dUygA_{B!=3\Ѹnn%#j%;$utshe)|C5m:Mu4uN[S ߛi;<7|4o\kO6zdTQ(Nh0^@q!N5JL4j2A rH# be}0z)#"b0) FрG!eLDaDcpLLqx2oJ'dx?|}TcM( La>grM$ ؉'}Wʢ4\+>md餰gJ !D-'I, ;\:9։yc;FxFHs1tpblac-GB҇2jU>Ji"FP`yA(yĀAydc"k/%%'%(e,xGI4 G&l&c \ԛҐOAw[XGWյ%6%se5״gxoɤ2CbH΢ZSy!,qVAX*F /6D_Fts8&kQO-tޘGE )g [F:l`9()((D4Ɠ})ԭV@˂ GPtVA+[A-*2~RӤ I"dfFYtZ #=1>~U<0 m UՕQY_~1 +Sc>Rj*`C-jK|KMOWiwJe@%QVU*&J,^ݽe.}dqn}B~dhi4-k^bmfe<>4g?Joܧ'n9q`GĎPvώ>3k> M=PLQM ;hق((|KS|)Q"jL$lT EVLT7l|h1Gߍt1PŘX H[X` H2J'fs?M)yz;V>ҕ>}=GbAOyBk PBlջ;nPu`@OnΥS[F UusO?tÊ,Oz,F|1vc*(Wi[ UJo ϰ2JmXHѧsp:CWsa0M-xXoCPغ %6ZdL)a#8]f@E7-7&,>Ã%:TBeaf)S`] ;ի'DZB)-b4Oh7%,BHdu`*| ;kpo\'ǵVXY uznck(SuUmnjbp=#LYػedz0bMjR&>`=*2jrM x!hcl4DqqMGk鄵l`V&gHx%JB2:Z=xݜA'b?VuE܇uz| (_8p{[+@T$3WW0_xqL&g׍ݪ=[ɲRxw0_VA*CT0"-KdZ[&ƨ52`d FgMtD@$vu7lC0ʔk5R"VQsJKb?]ߵy/e8${4 hV\^d~9bu_qz7* /hHfV)l7 Kr̽P$ KA"Q _]m9qEJ[=&g5d3) ae(2+sÕd4E\׃MkѠ-6(g^ɩdZDϱs20+^˙;#||9YplAQJujU+ `RZEdKZ( cB|= msRΒr%ئ`zWDXjS5U=X c{SsoMUz/%`LSь񫌪G|[nѤ ~c1f{(}Ov߾~j_fwowzW<{RJ;K\h?^xcޢi&M>z7i+ڽ>IANV~7qvo֭.A?utdx <ʋJ](oJsRY&;RBgq;"[ju86Y6u lN5RFRԷ޴bqAF_^2]0Q1'Y$d6ɡJg4>0|;K_7YGɚȄB9 ֮" 1.!H(e*aXejjc6&nYS~kc60.KQO#Or!qT}8R/=|~=Eŀ&/oA B9?b[lX]dk2ѬxLU>ցEd<2RV1 9-%Θ8<O+jn?5 7̩(}Χ /:c~b.3H0 p'El$CcX) Ն-"`ʒEy䤙?Mھ|/_rk*vŞq GoŲF&lڕPUh%};1 PHh"6EΓ*vq`m|~iʦ >L6)fJ3db.9Q iFNׄQߤ|̚?ٗ]D٤TyrrK([iu@(,`؛87HW/qK,Բ_,b{Ufb;ٺ1;y; ?fO"q3G#\ {%-TPc)GdWl MwC5BVBǨKfL]aNx,A[hR_\q2K&0ؾFǺP{`Wxe5BN5H|NdT2"d%BhDQmWKk)Z`.J`+:ir5DfL@!bHlT }0M51g{ED3"DrKuHgd!)jHIAɖ jM8'Dν""z)u1uRgT9&dJוР$3"& ?ȸ8kʵNFmqzE=​WBGR&1  YC&hQGB #1{C3 މb+_~9Fy~|&wam NJ~DIcn.@:F/~4Svݾ/(/~#oQC?dz|'.O~QGSR>`wVJ؋q7)R.^YYʈ VZ:pFxЗF _|3>Pib8ZHs5{~v3ϮO >wghw\;C_$f|g~h ,^atiy{kģqt<9ZwswW/ 66X\)pU3\U)=p JI& ,Z)pUEpR!\i酁 +XJ1pUWUZWUJzpJI+!/U֬=\U)•QJMbW,1pUUnSJqJ nWN8qXZ\NK+^/'/LHwX ߌ̽/gH( Ԏp[X=+~ӷ~n@X f6UgZ82ЗM;RΞMrAp~JPBZO˦a9G=.ڬvjBZίL7v\7y07>zΛ|mog={.ƣlQ34Xܷ5ӟ7]&ڑWkʛm? {U ug͵?y~GcڃN}OJQ( BhaMh5_'50}X~ LLy=2*(pXa'M4E/]8H%єiԫe0bQG"`"RSFDD $EV 1-ebj+'׫Pڧ,[e+؏ټz>99x:pUٝ%-~gӗ6y>}BI'E6?SZ('!D3 ?̲Υ+_90' a'!vg31&r$4>l|ǔPQJ+q4a9A~/ G,nS^z+@lK4uב4hq=dlx/VTp<\d8+ʠ@Dp,W`Oym,x01$z1IT{BC_ MOH=܅gI SYvL#Cdp(X}S$EiYdo7^z~q˪}냅W4|;4?OD knW|\|IÇI6{.\7]x㧃j6:߈3 :8~ _gs yʌ]=gF)`~/yQF.L1 _( v^Kcko:.w* fr9S,V6?_,; >8Zz,Mg^'T/cMx]FXbsZ~ wN/c`_aV8rzqC/N32Nk# ~gy5:x຃ > @`jжMݶ4X]qu8jJy)ZGokxtfVP0QESb 3FX4WSSۦ㹀++xtAd: ryZ^T"Zq/:>/=Xm18l9.?T9IַK]%7]rz;%۰ێzKa|]BED cmWR#"hRʀ55OPl8YP0EHMLcSB )&8Ipmoܯ޺eGv9iL2ܥ1Q2p0S_+<麧C?,mCIQ-( YJ<`THP*: FN0# /.Q Qŀ A2 oQc0Q<Wk._9a AS7)yfyJj~c.R ̒)AVl9q M|i\v!Th }62ɐpULL-HU};__-޼9]L?*a0 e:կJv!55RԦK8+ ӓk(ՄƜE<῵Yepv1@51`ӳfm^%ClϗINߟbqaKH@#1~H]Ða",oO>ƵlR4? ]]9LQ9*AGOkԮrIZZP040ԥ/+_H#|qqKN^ө=_MgQplǿO'?w'OuoO>}|_`fpEf&55y}0|h&C>g.&>1m -֚(=.~n? nv~INקfuϿR4sWAUP%\S.|-B`@oiRɍlXkuhaiIm솦*1SwqF߮@} v{O.ƾy4H)ʝNcTg) ߀WwwS}:pU S /A`ֶw8ftLHP`2Kۧ*1YKnӢ$JђdKk,,^) Y3$z$ O :g?$r()RV"ø Ktڙ{ũT^[A`yQ*ɖ̝SZ;s rFL(C0Cg>Rs;Ð,1mapfWoF!++?%1d 08evseY ˒-"[d%٦lxƣE*H9w.6*E)x&5 ug,;l-#|!f#ʏqt۵>j-v]?^ܿƅv= 5ƻ8}]i f.oqGz;~'FAohyqՕiczs sLؾ+;pRͽ -5EZhWΚnL!`R51NQPF]ƽ/G?/\4' ҈.߲+etvuswm@\Nƿ\D*j+t MSw|۲J?|ܥS'2+(2Ie!j;eE{oM~;^6K[s߷>6^\ X{U8*!rBf2D*Vk5MAJ>et <%ƭ $B3&H Ұ5UR&A)L DLM9؝ e,Z#5(O@b2=qW ~J<&К@SKa׃0X7CN>Tzvڴi>W2VDD8FQHAx#q4R੉u`\98Or]@2]2L|#ɡF$ }W֪B+7_?{Bί6t<#<[r'nkϢRC!}:9{|T]OFN䅄$LQ *RPW%P9JL2s:^W}U=wKlT>`+Rhe:%*E TI`4O)s^AOw.=i;zm> 3?VJP;m';L3,&W}>lYLя'!@mGMaя_Sq-^'_GT{'1OgI|=ÇQk~k#R.g32E8R߂GާVȹQvw5ћV7j?j_"n_7u6]6{>fo\gEƆz7zs4kTC?4ߙ]OF| àKn¥ aomo'Dlͽ oȍ ك,oi BrHeLH2   6:EJa9ۑRP)x .gK.XW,(mv5_a}EԷO~Ӡ齞 sĢV@LZ8$9@R%k9a ƺqIy[6I4dcFs \iS5&:ٜCc"^ҩKxL̃`Kqǡ6Q`7q~EGn= |/9_FѤ{?g:˗IR}9n>|_[y.#8,484'$L3bǂ1A< v4c@ݍDh6d_˛Ihλ\WZϽuÇwۛ77^:p}3?z&x]Tc=Ju1oNzo@%b'$rhOhB !(5} Jè~dNwKe>fo('+ȵ͛wDZףUovZZ@ >CDiR@ZS0h65nkY].GK/;CYOp)3Ԫ'>VR*Pļfa*m5*Iư\Ƅ ,y%A /Qu/7@”}@ր[|W2|J[io?.؞#tr~U@Y2%mR R6l{e/x;o.|o+$Th B*멱1GCDI`4(Q ^j݇&M|e J_i>*n[nwݻ.w϶ r#iӜ:t.mV6w{=7ӷHutu=ܺjZos4M3y6e->9wizh<Zn\cޅw X϶-K._Q v|Y6hGVbE7%8~fx' 2$*FYkW)PEt/B^XWUU(8dehQRJ[vgn;s݉,/,m]cC{/U=IfqjIf6#gkc?Fi׎o㮎؍Wݴ}ýir詓f+<o؟w77O01S{NLBk-n~_m]8c#3Sԭ7@ۆ{& GPtLNA+;Atm!0N5mؤ ɶ~10fŽyÒ囉(aG=hgnn1D-0Jxd+* 3d@Kb*'3 Q$MD#遹BS]vWMg! It4:lDQ3J\'/Ae餚./kYꐃas-C) %C)t0֜Me)"XkdYxr|yr@0Ξ`eVBHZ2D*V}گ>ס)H)S' } rB*` hK-u@&MJPא 񰼉l5oOnV3Piˡ_y]s^tv!N&JdȄhim#JChQk, BioѬ{29,*Kmš!HRL6ĤWܢP%OJ3-Fv ^g9ΙR ً>ԦSm& U][gY¢h xqpޕ6q$ٿҡ/coLuPlȒgX{k;_F  4HޘY 4HHj*t uueˬn| o# MF3J*UZ!I$*E=B铐nc4tR8˺؝x6A]=ϡ- ̒r[+"gΥs !H6ދWHx4jΚ½:YO|(b%TV^ Jҭ0\Pl8صKP0pxa)S'BjxNIj[𪀭9T'ɯ}dH} Rզn7!;Qzz'*)UdZR!E0ۈ }5r! yɌu`jWVS+¥m1`BL[c*h-µ* O+Vq$\@KeV8e*"e#{ RDhL$xXͷ iro*F^܎\ t ,T ))BsU  ~bkk'cO)c,:L;M{\і{UMh+ya78:o3ycfDh&t:2.ے#1wsEuh0w Msҿ:!VWEp0eRLg7u)x5(}:wqE E)R^a JZh`}ISD(N$rM 3a j0`H s ;jHwD@4:&dNsfXTqLfb Ayij}=V+?4%=I久u}hJ{OPc9*$ި":Lx ǻ,ܠw:% E {gy ON0rާL X?RX+ +HMM"Zφ!KψWgj05$K#l|34&ɾ1)|0ECI{ v=Y9VGY柱3H))"gt֗RLݙ)pig Y~6[  8PIwWIt] 2peTVQͪbv~ChQ)=u6RԳ3jR|ٛ]Nzv2Ny&X] m>]^q[HyoʕޘYU<ῥɻEep3\σQR}7K"Xj_"\bra KCK0A-1˦fHc3 f{DFRG VƏGDwm. c6JVlji\)d/~ n<-VijZ埯^TTȫ)-;S0)/r<=JWPú?=vsX/x el%V-ugl/cx2>H[ZSjBxNB㆚Ve\(mˮ$/ڶ9v3s  &| a^ghy sZD)3lnɋY;Ht$ O G>$r()RV"ø Ktڙ{ũT^[A`yۥ\Bzw C8:fRl7cV (3`} p)?_X P0?{}p Y(Z@@1O |M0&;tS{5]fb.ŷ۞Ip߷qvw;;ico*;+T;-_\u^}_2З:ކl8N{n.(AMCy~e \jAuӫͨd>_\db|Mc{% h;Oۋ0Z :5=l;LHKQ#{s|6QxI< 9!BhMne>^OAW,9ˣ1\~e}3_i #e+2*(pXa'M4Ym!N5JK9IN.12}Wmp`Cyʿ8/r!>?m>=M;~- /ӭفIaϔICBn2˂N''p& ?ȉJ F8FHk^Ɲb3(1&r$4>,|n[Xe"F0@l;;iYo; ^wzkK^^nnPJ%6WG40uxFy@=P|劇 9vZSApC4Y **A{ñT\!^0mNxمTO4צyT5,{nze]9a^4n6VUg Kv (ODx2/Gٿ].̸!6CY/`vqeQu6( 5fQ y=y=ˆ C0#_b`aT;K C],5 [>ՓWj<>doX}ST[e+.XNudV+z/,Pcn/(je|l7nIW໴$whCShciKo ?7Un멩Mr[^hc}nF·L8릋rr} 4uJη-]wݻ_>w.5"7ܣ*UPBFdП (OWѺ6jVw@y07hl3x~<ۙx;q<Vb#8neŰƚ/nnD=@wzSm>kbZnmޡmn2+g}RJWJ+R[!-=xJR2}%d;[$?R)y夏c8Q)OISC=%!Hke13#u1&3AY`x imr#OrpR2ub02 n~ûV4 .g &~x7RfY}vQkbBϽ fOYnOnSEu&c AϹ8TH#* @B#%3rK8rP[|m_|>M>minWYB0뗖lX'SadN6LaMGR&KG߰Ym0xZ1=9&W72œO!]s:7hPHv\RTq4÷d mJ9ƪ`B|FbooIn1dy`!DØ Q",2G,dU)lZs!KNZYqU6J(X{?d.qglͤ1} V832=?|Nzșغ9~ >ztDXZ#\qe1}L+1'u&;~ ??_fRbV#/~( &_e(I%hr+pmL%I#G\߿^rܰ؏Jb䢦0Y[u Ɇk9,q\_>A胲[F}v ڟmE^._v@[^z^<_ ˓bŋZ.fr269 0xσƑq9:=ٞZJрqp5,I?^ ?K:גeW|^5<(82rVd݆c⇴%9?l[C%畔–-펍|g~{H% fXdh;3RdՆ "a<#9hxZd'PEf%&zH%G1qsb)HϜ2Z$Sn唶E,ܕPri6xcynebŔҴzSֺȓsX f%Sѳh <KVAnevB[K"r!X[S28VI(dd=ɳ:169$Ia߹I˜ϝ ϵVHXE zJg)X6I!Ip(}P-ThB]J.8if"2օ]ȹ dDt>b2' Lu[谨\ՌQMFߍIƌFFQZc_R@6^3ڨ2r7b F: fkD^/c)c%1pB־ %h*A^hiKQ.|Xxҙ\E&Wk˘lZZ׹13{>Z J+t1MNL61altr\oVD1IO/zNw%qj(-FpI)[kUlF=˹SZ H! .nTh2ǘ|VҤ(Aa9RUE[E;[D$h.`X@ܗc 9>PcU<%!gQ1uVpz&BM#[M=? T|nbHRxA:bbI$&2ҲdT+ u+H;Gh}L?s"0ְ lW-Ys%xDNp6C0G{w" 3-Šu6EK{og*:ݐgI֌@qG@qN1Zt.8`G'He<h:#%x~C_;qV>padސux7βkr-YkpkzCsho1tju/..2m񡑗Y\Iv姗ٴ9]gOʡ\sr4H$Mt]Tm*C`皪xjC*q_TbF6>ő1[S8gԝe8y\'P^㶃*4:T2׉9'u^Y?/_;7 G:ý@BK l hiRfi,ި{lB?W>DɁVxԑ]Xd£ r> !HK) Lob^CNXugrm_p#ͱGGO=<::V0u!D#}$JVT9Y)` h*YPD{@" IȐufC@ȑ Hls2䴴.`F;Dou-t@H3uJ'^M3/vcMiO`uݽ 牠V>7^ʳ;?>L> ZI2-+njP΄ԃ/~ld1NģI8N'6ANYxu[))%J]Y>Y+JKa4D \+LQ CuFvD:q2F3>}'jwt?jTkUo2S~DZ2j!{6C ̥̽a;ddGxXJpCfQ"q셷ᝣoė;M/}v7ۯ>%We@7w޾yEpƲAq+I_aPx:@6l=ޔy,*ؐӨˡOڐ|R3`>;8򐸎>rd |H%z I` ɭKN%G+|E '$./E\Z8J4ʞPRkћxدSb-7uDQz HOC\ 棪R9Qi_w3*B2aU櫠"D%Ec}C6ƺ{Aa]Nkew 51dR W߼ލG-WK nIq_S c>y{G4֧QqdYګ@kdpذ|V=;={"'V_YW0l'; xϫ c!}U>8/9Jn kD&kL . \d^c:1{WwFΎI6C5ܭG 2_Gp ΓrLmeNd*`Dp2ɃmAŬ%7$GI][BGT sSDk`WlT<KqRXZ7Wi3>gUTEǼ&WuT>Z^Y] 9]۶Tt-2ۿj!w*`k$%է--aE#e΁Emf^,5rxs)K.H`,ǜ6eDL.gFZA&֌CΰJwYg{՛s]Bz]pIQ x˦vn57Mypqx'돬 R[PՉlc2d" A@/}1O8.Ny:-C1 &CR)F1d3>R1h]<"gQZ11vy{ㅑ^Xs+FF<'F5"jxk}nF K#hUmlWΕwr))%$ݽi  5"倘C7 п[ 6k.x" 4CȊ2aML,f489,Q3rڨ_D@`D?6ED1"{D\ksd۔_^F+xqODOAM(eb*()"Pd'$lJ3>`U2pɖP'H\1"vFN0DK U#\tlc\=.>xoΩļ䆀 2F;iyqK d=.;MPwj3<(VLLgCG*G{fq^ 9w [8ޏ(o]W(@P'$Hׅջ^wh_׫LSqAg0|َ`+wNc/Iڂ\~؎ZavK"c;*R4#)3Le-27k9Ձ;gwt:/c4x|c1inBAnChAq~T] 束6O~vB-8Lߙ]MF9 `Ij^{ ~٦mx@82ܐZ7]"oG %BBCB-}BBJQWDW =*Ѯ B;\*%+++` \jUҘBZjWDѮ\8*Jw*Tk+త,,gݻ^Za>‹ՁPO<%>߾n0M#;0┫S V*ĩpijv1jvL(;F%K_114ZET)ZeRAU0m3.X_k\1"l5A/"(l@Ki?/Pj`dXBsuŸ~xw4WY$Q)bUN'-E$7bۤfݯnE~rURJ_)BsUjGP }lnrjPkђ`iU`ઐPPKPWVxJ忎Ǘo~iÜԩG$RmNIʠnʚ*>!s" JyY1;2 Ĭm% [AϞ$ǐ/A8 k%v`arZɊ9;Y`#+/`^yͭ43WgTmBn2-Hka9*P]=]B_[9cQЬɳ<8Cl<ש"JE*bT%0}S5d< ?m1X@:4PcuYR ɻ#@OTe^!-D`J(hi": QG(YŤ.onV[eެnR&@ӣ:7=/BgWΧVMMNy"`2_o0̷=ݜ<|[Vz0mԸ͑amԫR:|vzU*:!ԦH;'Ohabܕ;\h25p6Ŗ)Huiw!;uhn}hFZC&蔯kd 2!唯P>4P )WyG#c;cc;,&Ma21lɕg`+fIv_g5f+hvTf#Z/X-ټL6^GaPڸc^x^je\Ed45]٣4S,i^6 Qg|m]}e_%r-iR|V.]w]z}oD\Lҕ+j]Qzu5o{NoڬO5gZVr}AϷ1VZr=?L5-/y;6<*l:Usmʛ.<Ś/x[jSm QN7|W:JDwn}Rosm[껱ԛBC[Zۇ~$}O&FRsdfȆgpq#$ 1-A%YdL%(itV)Kn}L.(0 4"9+sa'db].'_MinX h!d,8MhL/gv'N<seeѦ*<{lg_Y46@Jm-P*RC(*pIN,a'Pxaz`x`):^3$AYO Af2H<\dS`dOIbT2O_MQ9l^fP]Wo7<"ێ(&SvcIiV[CS 'CJ! V KܠO\N^K@qԐ{y°ٽP׆ ''oDgUqBŕgØ#GHeR%2 Τ~)DžjBf](Ezg)E][vi.>H_ċ7vA[a0IP M6F`,c@'ScD̉RjF `/—'N°F\x!M';{ZN K BEѲsNJudaR&)SN?uGM͟f3΋ Mbv~Cqmܐ.i㧳lE?U3Ê7C=w8;pd44ѕHyWl#Ϥq݃v/Zto<=y~U]Jij؀J|S D^~y [I[&{SmVv}hz9mޱ$xGSE~f>gU)ՎV_-nAWϦy[MiYX(HvG̦{/{g$Qf2'.Ҁ/k6jN m[#B 7=xVko7ZUAF7lpVpX[`|Mt{kq.r,ɒym<fb7w~Ïr>13j\&6hML{f#2eQX6L@3cMB .IlBŘ=zD9BHGBsKpg<-Yo@KI»W_l]5myyvEoYeI XA^scIe**eœ9Ro=co=h]3,tIޭ0&},FҳĬ`@f*h{rtA]A7b?kcʴyW?_@$azTLcv!{/]tx@d)xb~z+#kB9)k./vq}S6*i#l#ZV[ f EiQĒ#e,sk2XIYL)ɤ&z/"aDFziy趞A`S{슜0&k^6་BEZURNjVO_G<|EWt_t*i< Q'MO&A8iY26 Y0p<;A9O;ۭ0CBж,JjJzA%$4)>EձtFN˛*>I %ˍu Q@B4J<׀ I$8IRj'`Z$MzBWiZLHi>p] (Ml ٹٻ֟ƎdX|ٻRVIN(WeGU`6cɐ[}m6 t0SU.u6D͢r2F3=mǽMAi5&xm hfQ!p7bC %nΛ;D=M *{ Y]~˃V\ rZnZa| 'h0Τ-YgNlP|{Vm?CZr/p2?`npzk)p6Ǒk3?8j|ZjM_tEKa(kK*DH[9k7+,={ne['q뻹c>hS胲IҋBm:ozՙm#C-rY[d$ѳL"چ[0v̫ kt]绋gbŋZ<;h)]`O.8O)VN6{g{@88xIdd0G_#ɧšϞ*P&(m\8`YE/=YH]m^xHiCdZŏ4WLNd6hR9:;7+R)K7^T?t~%Wsy;/__Zh2o[Fk"(`3r#js, +"z/4b`޶ z}Q5 *UWAJ'lj@F"VуRy8thwC}k⥅>Q2 2hs) ,)ƒ NodcSƒZywIڈo&99SN}h022 D-J˲!QQTAWRwx^@\ώa0K?/y\:?>gߞH lño: (,L>yCTb:. ô?hypfmMOn>,WqȐa0ۃ#8VWVvo勛wwwf'WrnwB*_2ӽZѐoQrjxF{!m!0ٿR-(apeNփy9?[woݓf\gCt_s>Gx<<8V("Vr6k(zsF=A#IG:Y5X9:efyOa4nu`WA>|Ӌ1'ٚ))W*Q\5ꪹj+ [:-2&xJgS_8(x;|] =V:u~MxΆ O懷?Oxpofi R l ^=Gc{-/#CK3sϸ)q[v{k+@R~rQa8BIՠrF~kU<,QW]ylh8y-UF!sAU:ɭ q!zZ-'\k#]E_vąz1[SqYy;mE.%`toc `PEA ,'ufI^鬟_Cƍ:AΞGf1H)AM, XM{W^,sӊ! !:KLxP1d'!bLNRyBz{A("z!F'X~kc>41GQ#`b8ѫLx\Ztf伻(Li%̤o Шxt MrD냩=M0bL&7.]mPbq+Il c6-`]R*24i+87d[2Q; -1siBa6NfFAct`0Fb`7rQrd"FK󹆌/=^!tTΫ7zǏ֋zz\q1W=*f{"XW^έ6&|=X@$)1A6Q /N3\Al-ёό9m4a/# }Jj#\@D{IH"3HnJrVeK8>[+JKatV#{#nF:q2~%F>ݎvAS8!U{+ǟF-FZ*z7TLZ>p eްlu K"!2*+_iLwngM2xҾWPsj߬o7)̾!aaٴGM%8¤o9 oAELJD :Mk3oCʇa:堭1,Rsgd0gh齳s-Oߛ *. ̺Gᤵ^] 2M,<_;|z/_et4!w"7i4Nڡ &xϛP5yl"M 1iW7ZJꍫ@`6و@ -@ƤಀSI9`.01QKN"/ ӫQ8 Gi$'s" d9{&u46^'ȹg4ZA9>A3Y'1ΓrMh6eNl*Bp2ɣ$FLA/J*47$ht~׼VO" RK7QhԳF݂ fB&M$|SB{_F%}?q69oVtjV[E܉'g=~;z*x`7?ZCN 0ɄY5! }}ED#%bdIEeBoqsYx2J.H@*6"Q p#Z$֌sEaҏBՅU]pIQ ' xw|fՋ5h?v"@g$p.jfKأ$1$"x鑧=84{I$*Zb,QHQԦ46"E(5v`dɺkE;G1/֮ͫv #ʩ\ެ8 .'0J08r1 !0H)`QRt+ rEVt`Id` prg 1,Q 6aoܭ[~sVhQEՈU#F.ldv'NA;.GHBa6TRrX:ΘlcJrrXL19 \9 BY,W9\>zՑⴙ &:{㒇Eճ^U/V.B&E )v!SyKgi#VzqSaoP}{ևa TزڊΧ38i{%c Nݟs( {~dD Y]jKMt.5ѥ&DW BDR]jKMt.5ѥ&DR]jK=.5ѥ&DR]jKMt.UkDR]jKՈ5ѥ&DRbMt.5ѥ&DR]Ge$[F?~/CH+ԇoƓ\|*C۷_#=&{[蓷a i)/p2?`ퟜwcǣ_KGqϸui]bhsBCU6 7xm1w\)"vV7̵żvﺏo fu;kh,^d4. 0xAWlʌHjo@m۽büؖwčdIj) Izy2Cl)+ؓa8ю0Leyt騙Ҍ4Ze"㵙MNm$Q[PAבr%~x=LkHNtT32hφ,sR ^Gn3])xLg "K^v+ Ym'3IbƳF9Md5:G4s`Lt[^(A+DnD*@S2|Rڄ]`QD^,ZWe905ѭDP/?.}Ë%]wwߩkI ui(@h}ץ#7&2} n_Z7]rwtlxׯf!l"ݮ[=_4=_j&O_r1R;:Nƣs>õ(f] Xs1E_XwKmm`5/~UVm9Zm%iޡ?[kN?NzeFȢQA-B* "(RK}WҰMԿ0d!?d|| =yƯA g\Yphpd0R3ri,$zL%(ViT!K% r iRI.3 39LLbHdݔ6ҝf.䃌ބy)ln=j?=١x2Lfeo[~-QއNgNDƈ%<@K ,OB8X5“hs6*P rؘ53D?)g-AFDHFG׽*N4)geYY0lS BFz29IADϊ7rV ckjW^y=<݊kwWSTܕLji|FE+PEF6H.0|[ &gQR%.)lym̐EE$Rl3An|KBKmH@X%qMI:UI㟜)lHTriDdn}79AdKDЌf, 2$*FYq%R@-2<ϧ=h# 4r}c\mXo~ ބKxv]p`.:n'.>^jBg7 1;{b$'$L3b3+@=b v~q*) qOy&.m+۞˼[A>L;W{4kFQ3MX.Z̦MvYVVfxNS" %#\QeN`1 !Kb*g Q$MCp<jXW x&`Ggġ@,ѓhtVYDQ3r\'/o^#B/@Na]oMƮ?mngX<(6"6 zꃠlg .΃˯ ?:tH0N9WBTRA#n3ʶ4SpRZ\Hc!y3oZaDK^rA5[ KIl&2tIPDRBHψ1 qN&<5n#nt!_B2Y][a9So);hu6h!3ALnC̹sr%˺^C'VHXI#A$-ƿL{Kqs[~1 l8(]#hY)<7[lbqOLHF%m(EQ[&:ǹ@fIqj%7+I`%hKyAY <!'"*J`$9FƋs7^݋!  ts84)xiDΪQNnsq+N[W吮I@F90,7$*F+a4YGY^! u9iH#2CLpg dZULF@tY3mMB^9< ԠTjYB+ˑYr^(gN y gV(^('"tB;&+H|$Vj<5zJutIxH 3:Y` T[N.*7jaëowm1e(hoȵJYilR nT)R7\8:DT,T~>7-m Ͼ-{^{&?r2̀+A9́(A=90 8oC?τL# ,Tbw)M55BߣC2ϯWaECE4>|앗#3({fPo 7@ LqGx $ĄB.B#ǵ2RIMcF ryz0cJ+R%_pnM-Lel9g 'zsF fH Ne{JVҖ<]h'S&q:BQ@6Nt ;'{'ZqN߳a|~W*E|Ѣ_;?ߒcVT~YW.|O翏?>{tL9>;u307e$x0 ?׵3]ct-x49%fyovtoů_Oa}6;NZq',{G:Z-?~EERvg!1f@4l/Mö4)V:=ƾK1hO@b ^{l/GS4(PGtrLwUIY,Bӄ[p1FSԠte#mLsYmly^bGm ( xK#:s)>lLkA2'CF*Ickb|$4dSE AOnbysҸL6 4RRX\#l"`) te+Qs \6ؓ'xȚpsa'!Mp)P&!PTHj01:'z,NYˍ ZZ0V1/|2*u#θqRcQMrN 7s/XO\qza1urظ>uArK± ZvU:6=:BEGQ^>vq^nO6VxP%3W:pP h! O8}bHtdC"OuqX!=D9f@FlOJ)(GB0k)#X)D7׼zi# -gBA7<@ϸtqzG3@ĎO3OxSMg}}ãSk{,٣>OZy0rSo?0j }\j'Q^9X8B(S#eUmA=JϪ}9S<*Po~.PEᅼ~uh+9[ġ7q9zKMT ciw=b\nzmuS/?^tz&jr8n3AoX 8E]梎(rΟO 5`JހcoX=:;BI?kD1wJhH;TPuR <B Cl˳@#]gp!M$.9{y@ozXVp 5׋N˧x]?]tCXti#lQ!V{EqiJ >}m}g¢xϗ?oƤFͪzs;h afvF80l[C05Sqk`QM=hnU`eieR/L*VnBz![Z/u")rJD|)rB*` Θ= HÝY`7^0-d1qj\68x.JbVh <1& 4FΚg}J׫L;`VEMJ{1:۬Uy`%u? pCyd֖yЁ:4 jlX*M^ʄʘR6paMcH=m&IEJئ37Eg>,lhzog-|NQ)+V|͒xR@ ?«g׀^*ʔ"Z^:l(aBZuE^Wx#Dfpu!0b,C NPM_o)q ػ\RswY1PԮ&eZۍϞ ٧o&ɉt.J6ZkGm "L$AK/#fOgV1hlb _gahS9R@7Vy&}XJdu *6xT5:6$" CV::Vy#2Ao]VEU[ي#2XDrAEEm$E31QkIP'd7 !Y㍴шQSA%l Γ\<*Āj@1r8Ea6iQg3fꃱo5OJf17L^gALW B$ߡL0ȝ Bri+zB2Lm{P6qÀW:[N+n8֮L-.2akz5J`Е0 3 트BjcWHWP\bF=hq2^8Aq8Vﴼ1Qy qC1g8~FHK>U&;_? hwHL#ә\NwELgjv1m(0w U" kS ɕ|g`VWJCZq ŕ.L0N1$WqbUٽJqe$(\36L-mWHf2䊸,ufF]?{F9@;_!̞`gv70x)ښȒ"Ɏ-vnb["UX\ݻnحǯ3WKO ?WÞCZ)!B ,ITlgEs| r qZzЇ$5[ɠq*օse;:Z\ѓu\['UmƋZ'f5Z_ΪݱDRj\Vm.6U˪ejsY\Vm.6U{vfղd{ZlL-țrqd$bԟQP-of2VuG(fz4sS03J?~>x\!RJ+,3x,Əw}ͯ͝xɵ</Wca#nx'?nY歿Ypo!m ڛ[[z5M**GWܫY9*0T^y_TRA'кW[j^=Rq`8D)uħQލ4N!CMF Q$w` g|UQCW+;qfxsqퟹʮZ38KVGy* >yD iF-wRcwcm޽YO%Y ~-e/[=\/|zZ$'UZ^!-HiKRzܶ$9rjR95g(2obG$Q6m (ΝJhzϹ@a<ũ<@d9*XӖk녥!R"@]=pzb,'cJۼ 1.p/K ܩ 2fJ{k'o Nޮbw_x&>ߚz8IZ 4tL* !L MEKNiG#Q xLYY!. Z1,Uolآ3(T&FGmiDQXVgMrLBPumQijOHh8 dLoR]`_xMu2F8$(7L&%k 1\ K2R;踨\v5:&SZW[ZiJ( '.2Eqv}2vH1 W0jSڳޡ ;ҬoB[yb3ݽض{oSݰwg^ OU@@X Revr PJfjE{T߲kk]_y>y_pӹuc8tq0kcӠu@~,yR+Ai8AǣJ*OAͽJYp4DNӟ&NN# ;ƕ%A6Q䨊J|L *Jayk O2 ʞ4%F ^T) H'&duDGxsTcu]tn6Cņ&e}1u^!ܷg= % 'LiE;G;ۋi ="g0 >סk`I:?̻1?a=F財%-stAf'{Ap(|} xf;9$vUF:V7Ujȼq>3 K>B~WdKzB; =_j얗*7/۝to}w?s_~OO)3oopp+LFCF׽KMc{4-Kz:kJvyC7Gm7Z8M?뎛ٷa?yij6U!o~Wly0}EM_֨RvTX~@`_sy/[mU4qTGn(5KNV,u̖T5HCINҬ g;30z{m~ic$:'y+qAD,z}y^E-VU^ G|C+'^=qz H!ƞD5qB Dyfb!$lp)EIFrmb'eQH|sk#Dgpd$cq 5'BA/-y"裹$ѠO캳>1oټDep:[)6s!.k] [,[e3Jw>hvD>9ЕF{G훺kuvp tً B],9;|./qEPLt:A ~w}h|oʴn11o̞ז\Y>zYzmZZhW|M7i=23)+jaX=,|D@Q<}BYby]ȧO@C#8mbs1fQFqbZ4xfϠQ~[˯zd[uq?vXᄌ:2bܪa~1lςQK52j))˩?VeuxZ#|v`]y֖]9NP'Fq\{>/q%GeY*[A.,"طTY;~+3J+3F{eLo+k?ڔH{WjF<^\ =`Xn.U,B GuV`WM&F:Z ǞKo$A0ji?QK  5 yY ("LDH<a U[k4O=Q ؜d9UZۺn~dX|Z,>ö n9=іUDPjzDl$86ͯ ^FTfէk=Z'唳V!$Q bSqAjc?RG&-p-#$3,̙̫yT6Ġ |s}Lp@ AMM-H^Pl ɠ,N-GR^iÑ#܊g&*qiՋB};fIn, (FIbQ፤4jf$pv4^:ⲡ , % 2%͟rkG!ьKn5J,EnAQUw|fXҥ'LJJKqK]dC=Tk۞ xKWNxS Ƴ JZyiN20YV$\:tD#uWII'byHQlet1yO2&&N9P^d=II8 QQR֌٭ar#J͌{%t!'.|T]8(mr񆣼o`8LQ>{86R;a!iQD.*9 K6`m!z8.3ErDv~"#{YV\頄W:^&K[$)95㒉y48M["m3)nE 9'ę@ ( $vUv6x6I]FR2"C*Њ4_#+Nu0s"WV~DD]61IZ=uV@tF`t4:Fs\qjݍ]02MliWmﳙO^0057_sfDJi"F0@|Pǝ}0-(qjup$HaUr%gsn8 #oa霞u/¶b)h*_.&Wn/O`a}Y/.n1~ "Q)"7˧)?XυR #`iwխI)ƿ$E =~4ODzJJG_QGW7{{sȪS--N/+Eł0eW]d62=y* ULi9}$ɵ& & @ ]zչ4w>Hܲȶ+8V1+Ybam!4?̍aEE"J"e|Xn<FŨMi|2y$ㇶǯ(~45;1SWY-sU7K9Ӵbc PLBmZ.^k6 0|e{g&%e:tz =STlRg.%<gf[WM0xze+$)G>x&Rt8xXzqGDzxs}% Q=CЎid 5`u¼`*@2~z^@"'GO/8^GUǜ.;' ЇɄ&H03kՇ=6:O'>%bk/cy grgՐwg'Sc>}؉` )19=ONe؉ZE~mSwRc132(PV>0 ]Yye?!1I?׳*:ULHŸ DB% > n[]_M~ܖo=|6 T 5u9\o+Q5eS6Wj~AsPͧ|gv?^/}ӟG~jҊF`ocj,x97#i py:Ձ?G󣿟ܕT,NiȕTDF~W=ޛ `܄)74!z Q=wW4oKwYjM]vٖC,d}tH C!fH Ԇm@p)Oc.*5O61VքM+r=-EK9F O'Q.]^uњ C~t>=6lrsę8s^ҵǖ\IZU:QMLRWxI%0 q1xK5c}n%ZwYe{8}6|p5Fit}=:q=?z5ׯ5c{.RdXcfՓ?-rrݭNc/V!틖`6Kڻmޒ? Ӄޥt`R[}fo~C^Cp{пg"p=eQQF ;i(z!8HF*SAIN^:WB; JagJ !D-'Iaǝv.= OEՀo{A50o6[wLI *BG#A  hJa>UR_^%Q!jM҇h$YATT"&յ lInCUh<O>]zZg@/ yTr|Z  mL S~zSL2j%xޤ~e7/7:m/}O\z>|Pǝ}0-(qjupLe3V;%gsn8 #o[5뜞unHEW/w_ԸjFna4O܏7r?^$.Q]ssLErZ{["r=_V] t-!КoE_{/,#*&4>Izco`7J^*k⚀ ҥWkIs-+lRXUr% _:Bh:jQQon`Ȳ;d._2ָ,qQ1Sڇ&Ĵ;='q|-._0MMqJJ}5#r}t_*|f)ǻWlLjIMŋpM0R~sW҆&ӛO5lOLR]ӤlP9]=zM̥d:ߙy>NnC>q$ݽߨg޹7W/yrHWh@ytz]ʊ0~Ʈx+uve/&3=?8)m@ٺ4Ezc1>չKRʁ~O2 -42״'kay/:wN}H&D+ljMtxx0>9=QERA9ڏ+ECN?ojFԆl{2!8 oߚ~ѷO>Gƾ*ktdb)2Љ9'u 7G8uFfs t)bP; f2Y* -[J[;nӻ#S\բuwHՆ-I65& xKTڅڮC>6ZJ sRLWBp.=,=I!#a|[_n]CK= :rL9cuODž}FѺ"V~$JV{T9Y)B|#KYPDk@"7$0$Z8!QdHe^ȑ h $v cҺ$u-\WIJ% +nFE88"`ae2][5rCZ Z8['>)Eů9< ow硠Z̠A?Jɠ߄ӼD'g~|ȤRֶpdLMLS{V*ھ9>T)3ɬD.jML"M<0c!Z3󤌍8n%\ܣT!HCQ)\F_:Irde n<g=9(ovnqw{ům=F6Z,StPr۫wO_Z$*1TrUPAFb>0y;q=n<:-3<,Hޝv0&CH#,-2'%7SpF8[$ ]]O,떤QU6s7 ??}_wsغ9r >v &X^a*;y?MĶ7[הOvU:vQgOǾ.|O%߄noZk$5[t/ A* @#S\0N2%}Ѥѻ$]ZXk^+3Xv.̃0ZKNrJ~,e-;|8 lxN߽7/WjoM8Krz YjPs]⯂\\B74I(Qkn,C 4ff2NXFZH˨.NI91(愗.|KI%mN[V+\V|d֓TrD ٘\BiBЍXQKHe5*q?='/O,.LyUG*4-YKăӻ.uW_9d-DEi I$ 3`1g\ g&#/E*F`,}^#o7eH^q'gBPc,'DGGZ2Z$=&ێj.ش׍[ۍݮ.+<9 hKDE14.GKHXϳg*"]w* AIR7.)P"#(ɀ,Ҵk VHXe,.`Z[Ah"2 cDpHUe $׈ĭK`TYn`cq]H:Pe$o86$X q`dl2V*E;M.aښuB;x3+b]W^OOïC CqB/ go:킜v#B62ĘF'$5@6,ihxֵ8έĔOr8$k%yJKa>16p0E)WǶ[#g9"8+x1}s>&_<:IP^of5yr<+lu"sA3[b7M^2!|@yLe}U̪T g'dB6Ѩ`!&H6odA+09bfѺ+Ertwl;Cq0Tr޵q$e/]n y}OI1M$ط~3Cz!2 d˜aOuOUuA>J.1B A \%FDbWx'ߪT;-#BF8hϢB]"FČ GQu!@ +fpO/]O9D_G||Ÿd[ E^/3cE*wNtH ),Ӝk#1ˆRwEm(vp*s̯Yk/%7-MA\xGC(bٝU#@59AuH'.L}>ʳѐ"><S;L BGC>,Um{lOQS4Z{m=~<5ONiFNheZ֊9Mw-Ⱥl.^o;\)O:[5|BvrZRZ<7|?Yu^&}%M5v۫94 cn@N*IH 4xe SLX`h;$)r^7κ'm6c.2\v. lr6o`Z:z?zK g8ܜ1+AH71tCw<\0;PʑQNtnkh=W S l 48AAG\+,s鐔rRS B,JAwTz , īz?_/Yn1Lxa"K(«u~xcy 1YNR%hĂ52cdS`Wt 9b'_2Rw ̩4J)ca$H^Mb&> 2G\NhjboLd:_S?t1M}i>q4 t-!av6~_p}?VaO;r y[Mi44%I0$幅̂gӎy]׾׆_$,$t|h8fbkf)<4aOU֋J"rq)U@U!MHˢ$@ۥ1L;=;)ma-oS߿cw= " 8(nф^{PqC,Uѓ :+-SBTD׊dQE+2tVOh⍗觬]x+D{(I}U頂mgqmV'ѧWgEsA5%b_靯QǍTJ\z@ʺƧ枪JjJD& iX)w:cKYjv:-;hr3<ODSpsZ2%M2@4AH.>as Gt{+i9w7\yI٣+YN(@h"O(D$AHdZ02]b[kx՝U wt RTtwК*=$Wu& [j!-naIJ9bqO>`E ~7FSV?tI>FwVBZGdC8MJk< s-P豆(Eh悷<M$&g hI,|Q &sutaOoٔ͘5CwC{KJ$fCh}}OF2ǧl󛧏Ƀ-]ikkvqU`ۛK$G^s5?LͯO6 [QZwij|]3f6Z͚o]=#7͞,ri K8qrn_D,'2Ԍ1WskdmlHu۱)G)rWs.#|7.8?ٴD ;Yl{K@[Z?G)Pb'{ϱLsK<Ɗ:ǏԜ 53ΡZ:%CJSXOI '" 'qRM!@pZ %TD^\<9ur*tmAm3c+taCmf]MH_aDia4hB 0JY T&@)jXm!pZ5 ` 4Ƃ[`=m& /EJ\ 8%k/Vz*㘢u/=<(YYtмX\t1Tp,:k t6)N#G=z㑢Gmfx%.򮅶<βVE)Q3)p/F'(czfl $ TR$2$9PSj$9?TIEX*]:9mO Jc6VDL%o>'u{M̒&&d 7JXY4a&5zss6rȊj{7u ]QyQwKw/r]cG+Jȱŕd3%G v-zQ1.-m$Cd!@aa"byV|&y|{Ie2qdb=9%-auGo,nU=n B'*a%χz٥pebgx[Ѳ}-[r# LJVY(Qh0hIB2P6܂$,b: Ң;H 1jMx\NG4-l J["g5 ^KPo] %2%ơq61˛e>Л&)tƈAH I&N-$JB;/B$(n.FguYY!y&4JRW8sxOVD"FqC ^XV!~h߹tQ >Hc hePi*2H(irA.%Ppu&>|CB,yϜc Gb΃NT:G@VZ1$(۹W qQ\5QO|I FܖD;JJkrm87 G(6rm{/cIoB~Q dMWeVpU@;w yݸEvϽ7,DX=52>GK{ݴz_v[tm=P>wd\}/j=~NP<)V/ A˂L{D͜#iih)عݏl^4RjJQ&CRTF&l.: $OV`1B _o}넬g'Ͽx\n!*mKW =Xs)/yD.0}.j҅$>|`ָ$ g3]w ~G$ryٜ1_Tc 9 :nPR J1u-3k۹k ׭_%P0 8+:mqk ( UMw0 mwQjv//0x?_{>.1F FHtvv:9^ʁw2N}i&_50jҵ~x k,!1U a}6ӧYxr3 wōëхr{J̵N[c[quW;NeIK0@K[b|qK˚!Ka,m,oA} LV^s7u{sҥq'\&RqHjwӷ%)nW$ñQW|TO+?3׾v_^:ˋ닗^`.:x%: K.#AIc7v{Mo7M+c޸i&M:{=MUuvyIZ[m7Y[ }헗]s{P͞oɶOt`ҕ3?~1.k}sT PvTk_@b;E`_Uڍ[rbVm6.U#:[R^Fiv&0<;PlecGf y4H)ʝ䇥\3gn x^cd$X5%2(b@5d1B02Eq~ 1;2!y84[nBс;vB5>ZF@;R,.1˚MZ6ax{#.]'f#hc].UMb]i'W:`o4V F'Vrni0/pKg'4S"v^8)~8)é#/Nf6Po׿ۍ`NM~VJ<;Q+ўw̳XNc䢠L 0{V2j!g"8C$2pQ \(BZr_@ALpoZl< ܛͱMu)/o(vwxw63w%A_qMխՇIQ9>}:v:f3gͯ]c9N׎7>t`\g.A%ȉR4g#A- ^V8dnTT"I%U=_%r()RV"ø Ktڙ{ũT^[AZU{< JDLXڥ7*HҤVX﫣;FDZEQ}H+7gfRN^?~5(A!rF*gRbrW7JPzZP4b{QOp'T{(DH1`,&B[mH(U͵fd =!"=ARR%n@%0l\JPmWoLUn䩏} -OBW?^RN6pMn4H@Nwу;Љ DEQXY^+#ƒcEi(!j2U\6!aVEQGCJ9ō8m; ZDl h 0C5#4'(x0WkVk#gHQd[FBOr]g.r{u&i]H;-ܽQȦ(i tMp3O:Vb}Uń{Ú5u_锗25W2:e ,<`AE:<NcʠmvǼWUm SgmϿvYO G(wp)BSϘ8P=d:FY^1Tk%i|xYe)r3kEk+a~^ |o !*W Ṵ:~5"eo ί˦w\D_˃Vs!e\|Myg?1I+B[_߭\BfܯEQ(%H`oTQ0BD}/ɍ/U ]=bA|A5Ν,gHŠfu\*jT=9*>pG%?ҔJSpTrHk'b/c* exI "OQR`B<UX-$"﵌MlRVH;OΎT}HQbgx4(Jp ^8(K0YF$H@.(q+mb8 $PNbXXZ]u^Pe4s kQPO<:qtQq]qےZouHط;|{k`Ki?ZQy}PsI1/) b<eHXAi9eDH덨JF(yR>RLzKGm7 ,36r6(e]R1ZuqƖXXͭ )jpX8(mx4% r~@n:NȀi)% pr$SPD)c:b=.JC-FÐ= JlR!`V0ИѤUI*ǢEt!"'U݉um+&huqǡ6nP`Wq<T3R7(`<)#G=Sc( kYY1 4Cv4H$(@q 2G:HcjYQ?ΑHx(XlfD$ "6rcFȶHǂY-!Ɏ68¢kJbI5C?0BB2g($^2XobA% *$53ꎟ L\#:$_gm\-.q6*Z+l :kXx#5^pCcJcBЈ" .km׌l;<MNj^{s7bܨopj({i4dFQAnD?>Q#eZ#*)Ge(X ($j8 J ϰJtDp3\UCD \=C i%(ÈOyD.W}ڻUR!\1L)*+}w}&& XGIѯw;[C ИYKwQ66R*>ɾm oβҙRu$̩Ҧ{r\*mN^`܇y"~moquX4ϩMH]෪{3!PPX1eArEn]^v?p򱳬g/av#'' 8 ø.]R|@D ;8CK䣂[jb46yNFV#&OPҩp#Hd<Ϯ_ DqOexC+{U"GEZoUYY$m$h+l,fI|1s\ FUj9Ǣ%j"GW\v4>D-?x#8QY`jXc%:c0U|Z5ppY FOm >\[ V>QjèTWm I;"GW\J:\%*+ \="hNWbھ^r%\Јsfʵ!,Ҩun.SRi[) b,@U捍sUBYr%5(XIJf=@_O<ڢSkI&c*>C( Qs9qB3p:î?~}.[t V:0M|) HVS[NZ#2oDJd:]`m}V:4JmM"tY"cq4X"cT.dŞ.F #q7i;.]ʃT{ɱ6$Qs-9˝R 3<_|* C>at!\0s,o:nmœqCm.4Aje'Y\'}mnN7)W?<7K LFE}WMw+VIۯNs9V0#)-@1'HkerDʌmЖUmqYPkE7K Lִ ݫ{ty@Ubt.ɶzQ'ˇ|M6ú<<wЮrU zܙWx:;**' 9x/qAv!8g6yL4&30" -7plO<˶7L }-$I$A= #-q;Fݖ-)W2"5_$#BPmf78 >x!$f][o9s+EW6/E(` /ٗ|^ʒw`{gֵeK>$Nu#AN5JЫ:C\R5¯OvNiD7%%虒ϺM>қ_.^e;8ׯ/u\|@{Jdq+\@\v;L6E}H aMz7E`6qz*MQਖs֎凡Krg jG!_p/uAie̾}2oHۢe!/`f14=bhz/~@FhҴۡ@ieY:mʋAq:>j\Ђ,G Ԯ| ?јs\ҧJUzg`w췃àjԫ^wVT?ݯPpR2O'pv)v.#(?a|>O {W#{Zҕ+Ԭs㥔m^b, ;UfI{7A-\^ލm /RkYc=n}Z_:] Y޼J"+KRWl /\BW{;c,ۛ vU2Ff{gw'/N*#J11?~?8<\ty'鼼1{gbJoɺ=ݗ~pۃ/v3lKܫMkV#Bj֟/ ㆬNT}~aTT1/sk~DL×8 E >q|A!>9va{?X^)WǷd#Q{=^ Br?ǖ q璬5{ӞmMNWϑUl.uI%ƙ7KHE|軆v8il"v叽!tਸ਼?R']q)N',i"]]p=\s&LU5х:kUzsBʹRVM=Sy1v:[;cH͝ЁV Pcn5K wɤY;j͉!f-E֨ω6ZCVY*FZ0Q5mЌC׶WzNѦIcWzp]ARMRΕ; TMw JʵȎ)5z-pO#Pd0vcF4CcvW|D\L5CёBE5䔔5 L{}Չ&6fU݃6\6ʒ&S$s*0̿Oޟ7'!1NTR9=9ؐJJ uNF=]<']G[tjJ9o|0 46Q9|.`5 j}IX2Z|mŠ5@Ԩ*M|^YbX&UCb]ЇdTWOPc 37% !(R>I 0kjA 1Ȏ ўF Cm.5뎼0Q) _f E US`ّ,ܢ ^ TTlPtA[B Z 9xCyۦ CRܪC:}]1HmluXBs+}S`1גuV6 ]K ͑ձ{u dzE Sm<&q`T" ` HJƬG6 Y̡ m "pajGU((JU&ԓ)e ,9 [ʮtW jX9M fC2!I6q$8'XY\)tJි2j+$_gXnl&!LՖ! ΫmJ+Z1w|֕!ՠPwq+"87(c)(du Dak!$ ( ""*"=+FK7uk& A'b΂G ݄86KQ ` 5 đPI!01UR U r ֑lC@?V!jo\P;S )ti7 iY{TyGD)J6NjA!N) ƺuu3Rq'YFgU]%׭TB,f^ng $$2%FjeR":3nk em 4M4B@U+cŝ[ }FФi!:}p0(%m/fĥ*+f"cE-2( / 3s?` {blƱ;nXU-UZ01󃳽.0Am:Gfx:p4@.BϛБu\Ll#FIvNV*#d`18/;g=B1u7B}lzvBr޿vw Zo>mc~ti[Ӌ‹v:=;I~Ӎ"ŹN8޻|',w?lmߎ6mqu\pjH K:< ooM伹bwPFӛm#Ǘ^O?;Wq|voz~~r6?Q#/l|}l?/WȮ7ćжG>tx6ztZ/iL͖ FwO{jg W쾵ap7ahA(er0\91\{6FÕ=^h83\=C׈J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\Udd2Nn9+uz1+}`2E^ p%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp5C/p_Q]F W9Rp* W9wp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbՓ WRK2\0׫6ocQ^ 1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%Y W_: L._W?r4VV] mp[ V6U k%ra@.ea@T&}kC( ׋7R hzt5P+t ʠ. ~At,\kBW-x@cz=te9z2 +*.\kBWjdQW(M6n$"\rJb ds@\&dwZ5"(JmIm"GL01l>]tUK{bng JBF%'uutU3O+3~/zpxtp?3:l/!WU K"%?"_f0zss A0~.,BD@~~M#`FodxM#ZcNӈ[Н"M+R[ҝ¥+thy`D)z:E< !]! rB_qta]ZoUo)B3 8mQK0/ԁCpOWu bm]ՂKՑpB89]B+[AW];7c+ldg d|P[Ctho;]!JAz:Aby]`}rB=] ]qGt8D•+tŋ#NWȞNLv]+Dke P z:EJ28uj;MQ5  Q[.E:z-cG6p62.+{wmFD J"pI687,$у2;oq]? W`gXgw ]/' ʣ}?< [Y~jϖ?",{GZ;d߃[ 5ltRp0M6M$8<BBZТC-(YUp?Y8@]a~=^#L*ݔ= ݃o$PZ-0_ sȿ~X1GEφz㞊׵֔WjvAo)ORMꐩ[Bv(pꊩhj(v*L&Q16 saZCL[QQa smPR >>ruXmpW 4Ȧi8ff.rW";"Zg%bR]Pؔgo;k -^;dW   |3ȯƳVnrsĒP2{`,~{;ksD"*O#<8Nc`GL6%y#E~γIX7c})ZOdd"DꄴхR/YEd//C8e#4/Nu?@ [K(eE9b,82 `QixTiH#ልXtܰ\D)ޝ•YE6m- B>p:l0mBˋ NWҲ~)ҕa.- Do)BFt-#Ctute֒@lUg rҙ DKm QN9i/ˣoV.=rZve1P%\t{zSΕ+llg 79!BT%S=] ]1`+N;DW`-et(9se .;CWWvƺBʶS+!o1]. ]![WqOWCWR(E +;CWWu&ԎhMcWR4^\BEKq8m%Kagxۼz[zlA>v1rPll/* U_W[qߢvs`87͢1$4^2fR17RGD%6Ӕ'A r^w¥Fy+}`qbM{ 0?egsG ;z|m>I s<|"gJ(G9&zd1r/n97m&nlض)R^gQhR|OVF s|`]<)$Yv5( գv}uX˺; ].|4L Oϭ|;^r-%9Vn|V=g [4*AxZV&ކfVhaR0 +o͈)'$_>g\пIU&T>؀dl@i*K"aT% D26&3ǽXz7ϒ16]`;_koF5<:fm鎅y^qq&a"š&ĚHa*v$a EχjcxP¤Z߷^hL!&1ԆĻ8I,"PN)O=$<O鄩ݑ<0*aGyMU1TI)Y }2Zۮ>vMQ&6r٬\[*{odzԶzn$oEMr}r>l(^٥彃Ag*,]4_5ze!?.Eok!.ڏUCwN'*ݔ]q7 d%RZv_>ӕj/,S:ztwS2Mӧz?QqM*@ڌ vhѱxJ4__'Z(/_zrݕȚrJ7PˆT!ejHURJ=Ϭ=rNXs$wW<[̰;=izno~$\iUɧl7nmct xܧ`Kxg$DNCib奔I=g{3Oǐ`y6~yADz.ORObTTSCӘk!DrKt5qB,i42KWJpB"ć8^[.J) @') G&s`db.ͳϷŤԢ Qф@8[gWX<Ιh}5o-:NK~)?'A;B@8li"ɬG$"8s?IzFxF@Y.Ÿ'@ 4& ƔDeAA0Z1O7>:Q9{*% W{CJ2Ls3_OoD5iJ Ly-pY=0iv 9?Cވ>zcwh4B$RI eLs& e8#7Y0KE*F`,k#2XDdWEfDx-J_DdN-ȹ}mYRfK3^´[c~aMBVqGCriַe'Z&+ch\ph7Yʇ1xWWV Qj'-"%,5%hDHh5B%p&9KT#mrH~pI4DPY(5$Ɠ JfY @𠕾ORlKJ,b*=hαN3ym]e9Dp–x=9lqL>ĭârcjn7:$3%khIވqlI(4"g:; a,ۇ9 uhpuƟU8.y<1]otFXE[6(_x|l+|VJ㷗[χ/ 'oß^]p+rG\߿vԊ/d`9Y HLB u=[{ۮMdiR֠VUX-a\UvtgW 9z4R ʈ NG [q pvԷ&-꫗%^z %B`Q1uVpz&tzM ޭӴAu?6ŐtCIL ekYLf W:A+HTPUPZoTC}0,Ͽ\8.ijV#k$( NzO 3pP˾.7a_|Q|0R_(ίLnߞg˛InYe:⎌Kg\2qtNxc</G )#C36^gf<20d@Ef&YvEnB rq"Lzܼ[_>a?K·dJ/=D/^<_<_!QB T'8?Ԍ*3SK+Ni j9Gwksݛ^"]ʛ«l1D1 NNk+jmEj_/'ק~X>^tY3-ڑ^=aaԴba- bQeN>z|5))׎*ݣ'nusVu\G\F/Q,ʽ{Ώxo8=?qK:㡆xӇ {ϯzs?s_o^ѮZ F֑jk~ێm-wˡiҩ kz՜qo5hbomHw_/?Ty16X]bW&Ygtu=H4Jo:c:-UA!sEU p)VyEP]JJKW KKsV8 E} W^^ǀ]z~C~Koo~h}b$WH`R\M@PNԼ*졹.^r6h[Ϭ~9bD2+Rȉ DdM@@s@Ĭ1M7ں8 y^:G[/S!9`{e@iN$ST1;@n BeT+B4>nMˢ嚖cwM+Aҧ\Z0t;t'tQ{ttv1)3n|օiH*isQlR$IHHGDkgVE>d.X敎 @b0f! ,%u Hk]IR(cbYU 7YeB"s\2kwF:f0ҝ5k<;@Ak{BA7 f8Igd@KLh/I FYȬRB@ܔ>@vhU__cŷ69hV*7P{goM|Ү ?-|UvOzNqЛ}eD;B62u]6-r7#@^c~04X*jWQނxvZhnoҲBixJR2Sukdwj)Lu5BC)oؔmkTz̃.pRSjq}ѲdNZ+RS19 MRPz/x Εt OV캥>nD'KԋN<ҹjcjtzσz6rÿUjL248.Zk`XH,'3󤌍8[H3c 4m;i`ȁ =R$t4ғP| ȹ_W@G$_\{C!Vay]Y{,7~rg/ G!I UcM0!4p;뱳ztCxrJ$N;o!J$p̒)Zz#L*k; #uܔ ??~O:zø >:' H4k/OBY+IqB/ o go:u;>CcNGpĝ%$6,s`0fD2VeibJbfUYqB[0 V+cmU;V]hGj~h<nc~<׼LZP{32"8d,"@<,TFfQx#D oᝡ(ͦ=V|5/rOdS~ ^Zo;f?c4O]ݾiUf3-62G259HzW(`QQ*o:Sf7~tۏ.؎Ѳd%C2l .*X6da ){ ud0M(T !{;.p$ڸC0!u@{S4`\ "S|@@QmXjQ9|nT[|wUe_(%zc#8ჱ++^S:ALWW\?v`tڽr>vDZ:Zr^ u^v`g; l>mGtR9* IzoN*:"d*͠ :@!*,Nf'ؼΔ9s8R51dRG $9sLNFKgeN9:oswp|}{ҵrP4ϣ ֐Y7׮ٲ-Nw,SɔP__OSh a kh771f( c4чs]~ RZ\EÉlDKNdƤtI`l@TBܑ$Ȑ0U*#c.8)(N#> .2K l4ρH񠃱&1[UwC=i6oQ7aIئPΝ<#mӅlٰ2*ah$KoM ;R7^ݥm''8,79vϓ9Y{MOfnU|$,2 }iq О)ڢ-.I#e΁Emf cugsJ9V<ƹ%z0 2<&FzεsfUu[5td.ܩ.\PT6,xMb8J8^ g&ۉ<Aj`:- M^2!|@yLe}C޵6rؼ/@?N$&j{mKnI} }G-ۭeʖg:KԞpq΄k|ոd[\ʸ({\qqDOB;y5Add;JrX伄饄L2wոc[N.:o]YMdt]Jߏ~D|/ŗÍ/mZ[зV$Q:~3;Ϫ_&ne^V}#+fLG~ni,)6iNdy.Afo&̓}󍣙0iSV $/nة 4??S>긥ߺ:7' M (\8Пπf@H)x;F5cswZ"I!?;)r,4z lwe힨Ւm$es=S R /qg,+jZ%]U+T5jrpyNҫ]rE;N6?.uko0z["f}3NM#%-rZd0x[0 ̛͈,8?^ uތ ᗛ%uhMiha>7AI[H?ہsw:'(,vL R OC;]_槍Ku@r%Ô?g7٩a9g$}[fP:U%v5jf|q~}M2W!m:IQwG6$Z,hZTBN=i1"XmWyO#Ɓ֗N+~J79+Y ,YU6DF(LȒ̅ g!FiۍZ'^"<7IqÄOYN=x?ܬ fJ)} u.T+v9`/RZvHqׅ`)*y(BB! "F}HG`WDbsMP+P)YW,mThT"8]CB ]}$ 71{zEpEoi UI'ih\ؒvFgi7ÒZaYfKDѯwl_3xa,N%@FCR;4>Ow4Au8^7??כ3\V.'<=Z7|bٟBXڢ,5=8zMqZ9ԂK0NN+ds Zk/A)߀3}ץϛ. f<>zwdnO8{'k7?m?~+T KYLܷiBrRVs& *aO)Ptcr_;9ʭFVV֒1[Gĝ*Iº,pqd OiurY#Y$Hd`rH .1s))[9s&J=*x} ONEyf†rSlIB;'X\/1"E]FMv%31s,M,TBa!$k^x, dFE/dM *%(\[X [mpE@ 2+ntNCGdؔ dj}5`HH"2mܵNy`M<q o5c;*x-g=ODW u^HTYiTHj ڵY(pyEزz-nv+4DR90\}m[l$~n N|TZWLE%W_H_㷉~6Ihx>1o*= :Vi|>5ݖ:c.e,0r~z6_Km R2epȓ!9sYF ȸVR}q涬 ɝ!\ogi!>4\.giL]?ӞSW=iMa{_ӵep1~,)GRh׷e EMVn^#)ZTe Gz0oh%(s#Fyg75V<2:Ywשr.ްr8e5ٴ-"7{UT= z8TPv!jan=dF^ۑr(tT}sTJ?ʭhTU^ L6y(C&fD$@$RKUOZeacJo"<ȉ .2Mhr=Pym}b쨬FzGe[ {ÜVޗۥ,ɏ[ EPέV_vsV-s.j%5snW+~%UCQL",f^*r3YVJ(L$#0k)s,@6 Y֍x1K.zR(&f"mLk8W2X**m+ʕY;;'f$"ss<:ɟ6h?3b[#Ke`FE83嚎I&2f %fG^ 4 sQ0齐&hHch2vbD1!eY~8 &hc_Pk8^h7$/s6AesTrbYx]u*^; ^ߦ6.x8L!L&&Hd3jR cV2q2V#g=vFD}Aj-"ʈ(zDqhɶt A:=Ŏ>9P 1X7~rƴ1ke4 r}tdI`3B p?] .ڝ%ZjO8ngµOj\-.Be\=. '|r흃ļ䚀 2Z9i,r^RB&s}jܱ-x &Ԓ7_Ǔtkq W[fK!a]Ap}KE?ri)6T&rP5\{RB%5^cRHTAkClIEr5έgƶ%hʮ~Q+ PTvl/EKlmqYl (yy)Š$CQ{N ʘ>*vTw'絴>=ft=ۅ7ﺞȕKގIvn[uŽ Ks}ԺKvzu4Noڬ\ճuH-[j9?ؼU;+r5?͟z;Mx~8;9|Mdz:m2¨6kXs>y՝Kr,NB{szE 9/a,ޕBDåqi 7P5:c&]+h!zSR{K}7Dl-GљOi8^^n<`砾g")(F#s6y6`\vhBƘJ1e.Mny@V3!@xi*Cʒ[2 t5!Hmd:0\fX3Q =#B{=~mJtC|ː|:z2oZk8?!d^:zyw7JHҙ<}Ɛ`g,jLJJ % J~)z@Orbo|;A36ؘHOIژڊ 3A$.ƤerKR,e@,w%4&M>m  ,A-ӗSTܕ/x/ <"'x@Z(Sv^ cI)nq:>(]9c64SL =V%!Z{:\k D}0ך[Q>~Pݝjki0$$3Ǐ['޵6n$_vz008$E&O[Yr&)Y%Y(8AflTݬj׸w5>JH^^ Hm{Q=јyO\,cU|jQQ_j50b;bzX>3YRYl|t8rU_7 P_4'>6xS-wOp Ը=CLŅ`w|ve= btu!UkfGBV Fgp6852o~ZzƥT'7_7<3gNnC>syfT#7[9coU7ucOZ#05`d 4Iff*WHR|LT>0DZ(7,"j?iv.BhO9c"C`0:aI002?0$NB'Z: ]۾\2l!sy߱ONG]Xb3Gdn9!J 4*wf٧nFlO²4c@GgRPA%Wc BGUsT;ߝ l0>Ūnɧ?*p*0;Pĩ\XH(o%c/2w^[K?ӕV& 9k'vKgvL Z:{)g0İcrp:r!Id ,nէs3|%gyJkxЄ.2dk"m2JNTR"xNO 1:Uu([cR;ĬW逰Ied~rVk/$5op?M t0>_A6 ߇¯eA.x}үzmq07 *L/^WGx+R B}Ӯ*2\HKiT{=Js0u'LŮ7=,g?jiq4#iZ4y2q0K;_ӃghkPdJN \).O/JX)vʹ]ܦVs&."AyI#FxS{\zEtީ.GrD)EMPހ6%lt>7uUcϱ 7Q$ Ӡrj4iEBjJ23!J֛i#cF1ZleF4CPIwT!S 3xq*Wd}m]WuMQr]mWt,K`)[U64KI7§cbxUVOyR3C |u7b1I)tv8|nh/ڙy7Na@&dEL,<+{p&-Wʮ 7|wxSp#\[6C[r-=pAw30\rXpCvw:Ϫ o iV~,dz춸 L$sh )p8(@{RB 1:PA<*?_.^1Q#Ư0`K ή s˙B(\ǀ9fX:gW] uv) =wŕQhVYh H$3ǩ Z.3`CwWGi=ҝ< b]n>V>:Hq:e_ɛi.f*GqmR*V2r̥`~msʨkuzl>;F3alv*e6}vBm?֝5apiapIMdiBq3hA[uȱZʄ2a^!X&HSE57V"FMsFV;|rIZ. K$~!Ol`e mgɄH] ,;Ϯ& bqfIrbF#gΥ Ƥ٧qH7M߯6¿1jcw+7YO|*(b%T# Z PQ#p걫ֶCQ{0rHRh=>(yA)ıH1? ຣ5KBXUkdmw}/'$=p'Ǫ˳YO>n⾝rCzWoTXJ<`T@A("#*Nip4Kf 6kTs (WK b OƨU1h(<C 61^#j+~4i):p04 m4VR:RK SB 6,d53Le_P`u:NIH :q`UOkTH%1`o"Xʵv5 |uըIsL(KYߕJ!0:e 3Qh`|x‚ xdN79FʝwvQ D\ ܁GrJI؏;|Z#1w挋T)(a`"V^goj?C&P>UqNP`̂rѧV'WAC9Z$`)O M:N}wa' &x-La}꣔ybADbLTJ"'ۤx&,!X ,`!zGAHF$Xɴ/ J3L 8APfd`sc0ؓuǑhJ{OPc9*$ިB ws[ŅE%Rx/1 E͌QQpq1,)"F6Zu扴MvCPa,U\9CበQ1Uw"9xۑJ@ge [nhC6s߃-x\>#B IJH;Xs)/yD$~.xF$k|`ָ4Ʒƅ'8wcJ{0|#xe(bg +0s d"gJ<ԝ8׳H,u^] dE-6`M@ZTa9\%\.Pe0$3 ScݚHQ}3_Onn_/~0 e:%:sMξYÈyF)װ*UHgTݝ'7#Z~)+.|?\W\,*gK;07KorlGȽ6 .+ZZ[Y Y Cde3&aQ0iJ>z|f7Y9FV ZꬓZ]W.Ihul9߇#z:Yg8ɏx,}kTLV~YoP^\/ŏަ?|KL}_`&dƫ h3_vZF`ǔR,.3!c4'u;ldz "fσ[c'T5J)Ҵ\رNzGڳx;2 #~UfM*pT((ʹ"0O8K P (nM^ԃ``8ywʼU>oj3'~|F>^؜@jbb0tHמS7-V姜 M:o|Jbi%(^^ͭ6 HVT+~t1/" .=%VgjzH_!>NKAa` bˇC2IiG6EjqfT$Uw 3ꗙ/ʨ BͰnؙIAxScYz҇b=oY;ȿusHlse9MmlI)V-|sݝ/.foyf|6Q ,icYB _dtjҦHI)-goL).i[O\QlbVap-vHmr.nn)tcEJW/b'`/bJzR3+WЕ{лWP_o p26tCyl5nr+]=u,{MK+Ekӕt~HWDĝJrj 7.mtNW2%ҕCZ`oi1tp] ])Ze䕮^ ]1҂ Е QW@L8vR6tJȳ]`+kԮhJQzK|8 S+ K&+PnO8ZT5 >;zV,+M@,pbTe֘_SHW/I])upb h9fV*YsJRQBWcJQ]CWȡWƙ]kMW{N~pkW=CGx╮:%f=ZC{t(^ЕuY])/W --c+EiJW/KlhAtbbJY GOWmWz9t1 R/'])J+]@K2,..٥ЕJQۍsV>tىn~|1z)WƉd,M|=9{5&ܹȑ?ym}=##kϤc:ٚ3ZZ-,*(`?v1J^WB_BiI+CW@?vRvt* $urJ&HJQK$>~ϥ+,)R~1Jƣ(]CWȡWj/;~ .o: ?=ClJ+YCOV' R~pCӕK?fٙW%2 z1?\[ &"<2ݓ!_׷y6(yK4.nQY?B(/?3u fClz?E׹>>@ymޢq|su}q`K޳nn۴!βzvo]Cgן'>⛣oN/^o[);ŷ&~? ;6jOb\5'BHm?3w%od'īݛ ##PQü|Ey:{?[&qiQ>o??f|`0} 6CCCr{ۿ?nۈ. _M^_Ol[_竷ي?-gqĚG']B%wt*S2|+ex+Ơ||d8me eoo@]\n޷1o{%7SRg[sE(qLaj'}l%$[raݙJ t!8oX fksv8؈G0?QlH!~ZhP|mC! L˜D` mP($[E3C$ZNDzϽ#Bm4DJ7+b01whFKa(9PD.%rx{HXCbC $@IInH sGCLf sH }JcEĕ1Zrokπ@D{&S h5J{Іd6C4ن06[*e`031M~B11ИUr#Q>K5~kb1J!yD!8<n@#L'|SpkuM 1WUZ:NQ^dNw!ZICާ$YU)^{r}ϙTR35.E3ɍD%[C5yzs_saM$)w9$5nj9o3Z|1 <6yu!ɜ`/DmMs/*9u+)LIO͚E6-H} 9Y=<\hc?cYB Qd|n` :`֞%C"XdGhϝ`Gh٥ ә](P!}b0`9g0أYԡC[BZ 4Bs8?ҀO+(9Ḁ7C/iW"Hmqpl5K)[:+B2fCvXv4giF] C@prIFҼ1 [-`jPѶ"jQB-9V GNyA)9x X!J ] W cc W؄&R) d>࠸T e@iZV T77"<(n3$Xr2yA Av%'= 5FnTz! 2Pư)0m2JmGV,JLh)D<&ʄ>zRaGqC| $$6Aa-͚%@rej6FcCX D +) XV̪%#h)Jx7Fa b8#,.@GF]lc:LL](- A5qwpD6x,YeX ⬰)flk'jj@2Z?~XY{UdY5լ›ڀJ5$z魙+ ."1 21߬vC6H@> ߦ%@% CzMm-ٗi7zu?2\$K90tt!eO`3=4`="HuZ:5WZsLdG0癐Fh Hfrzӣ! Mʌ4MaRC^"$SiChJyxHsX7ec_ >,`E^H"u)A -$\F1 ϺCE7 ^VmQHt(5#C^:H9-йgǺ[cx hN[hcs3VjzRfzPi@H5(LÁ8#Zxv%Yy~58ڰ1Y]>@zM[p^qi` EHN֌v̮ZHj䋑au!~Z)AH@.hVikzm(utqHZw4 a4 15fs˥:pU].I!b!c-ٱuPz뢊H 8irs't\5KWCq>J`j?6nҍy?.-7[{äG° 261Ȩ_K3黿#S@'.n!<zӳ˛?N|/o$Z?^BnҮ\wt?4՗Z_]~ Fo=͗s5d+4A4sn+xdLWwO_wo,{ztzMc!dɧ|~6nd[[ 3y]٘N-23*0-a9@Sh}%PzZ+^b%XZ+JhZ+JhZ+JhZ+JhZ+JhZ+JhZ+JhZ+JhZ+JhZ+JhZ+JhZ+J\ mnHOK T|ɹrwQ,Kr*׼8A%)ˮY,@@\u% zf{v}LL Xa) @ %^a:@˚ @B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 z@Zb ss&0l9?x&Hi 2^!L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&aO[\هOjZFݤק6wH)ֳ^_.w'0-TVC\*꣹¨Hk qUz$I\Zڣ"WEZcx Czp%D\;*Jq,pUHiWgXUب+r,pUe )9ի+SrY~)>sSԀW{7╀rUɸ=] -6a"tD3@%%ӏoMǗfs/(遠wN)[ͷw*I*#oY\Y{sR@4*%B6Q$L3b`1`V1fifSd ;yV>k9ͥKl.EZf}s)R t_K>"-3GWE\ɏ:\iVR^!\&Q%?s"<H޺*RZp ᪤/yyԛhP_؛i]6oNnڇqףf,?6T&<׿L1+OeJɂdW^T%Ev́g,P~  XȢᅩ_I]eO_oN u 6uq OsX zDŽtJuDc);uPgUx!͓E[9R[Sj7'7e}R9iޜ0 Na:l.mFfw[zuw5j6VzV>+x*n|Ni׳J@\VȦ䢏ӷ&.K&nʖDp(b53"B#TInWƒ~n~}?_#d J*38k zdl埼Ӹ=uz -[ѯëOg5SX`NC,ݨ[)Cp6| ̿㶦pɚ^vb{qjl}PE(I$N+b2,zF 0ͤN:9{R %ebyȜw]Te;adV ;ӌn½bB2˫bW4{؜ZTq0mzTzW؆y\j'l`(‚3^͒*; 6`OԵ/:Elnx0 s=&W:KXL6 163XXKMLjݙ8#⒋y0lWq(MтݤLq+b(WZ~,( 9b%̃RւRiT ! 4O&&K3*0c~$cvc VjR%5BWQ]N'A2ZR*AZ NbeU09,Nb2A˛R쮏@ŝH{}ĻnmÃ#-m$[)T/dRYa3Dӏt`,$mr&+OK6[K9{!N3'Ntٔ&D T.Pgχ1h09r@~F>)ntw]Fȍ1]骉+KM} b"C+h݌f7w9tfsv8gZVr~d}^7eoR077@͘rmo-϶:.}=ʢ,8|;DWw mm"|!}U 熾y7Wx6(oS=x-pk&y,wg1) %gx&q.qd3qLL d:K)Ri9ҚhB۠}\m'}xOu,;~lz,iFgL ZyB@lJdVh#HAB9;2ĀxD BV^hzPcʢDYXA>q^x'1)1Tv )BBL"X+ F989S|L))β10t&0#FN}(,pE$x|^[S`x8^l`àFk3+SI+ #uLiYlIL%]aY⨳*2>O73^^DЃ! r)J*t>RedK>UYßRp$hHT< O1]?{DD@d&Y++`ʭ`E1Gji"dujn)VM`z۴Xf?01P2a)> ّ64]OAacSMu04Y2A]r مKi} ZkVraU7MywW8z/ǓI@Z<5=SS{l@ozy:;\^\0t1 rVͻ] tzd"rRu>^kY N ks?/㺫5D5BnG GOX6f}/UVy ʦ7{[Hf]GGCR-F%8֥`>&Y6W ,[WCC +lRgUz?7LL &ŞEhݶMDۭ;ysLR[ ,R\-LV-RΠMܞ-z=STҡ%fţuŒ\"*Q]ͷlett+þOk3%H4yDI3|A+yU˫lJ7nƒǪCͼ;Rg"xXem2?y%2;!471QNVY7'%gW0%xi<45߻2F\M5=IZ~?&%S؟a]4)Anf5[-q77ғۖ 4T N> ٪6͸]vyMbZ³ڛA5Т͸zyF7ћu~6?hW`O{˫騞@߾"WSWF{=h{nRhcju2_.O{9o'|«R*TJ 01"T ]N6HME= LWNY^1fFY2;B J %zhv=,I H>s IJsZEkeĀK[mX i PR1yooVGM靰:/1֋B5s9D)c3(xxp+Dp}R𠢷؅F#Wנ;S85l Zta^ȅQD.<_y5;R~KCRxVDVN_&K. .> ~JhCw*'{'Y{/;;qń,'p2ܩ̈ X,NHC%e6:AI kU.?{Ƒ@/fw?dl@=X;؇C5Iʲ}~gxH&Giܺ(8>!"9}J`і 0'{]<92%=Ԗ')>Fטެ^&͓o>Q%M2`L9גJYnHT&y8 W*mi$ha.FgMk(4LjbY[*D&A#8tZi 5dV y+|q)>(@NR$Ê=qB>Rlb fg*>=pʱF#RISc0SMK@bZTcIĴ~/*^W5cϕv}ZcP>{4'#J+e!1٤@ܨRސqtXC6MN&խ=Zc@ch@::s3 pPc̹ 8%gB&SBTBXhw)M3Bߣ.nsyeB]R|t;/h| N'핗#V$Bz0:F !*. B' kccƳX&tOU܋uudB`s˸!kUA傶^) Tge/ Z$֚('-EEd>~ca9|ɠ$$44xǍ4>S<1a%l'IPHp\AIUY?7?TY08fJ&]10ޣ6=E,M9"8UN&60gE~(/ѕlFC?S۞p'uCnR&񾸥Lم8l"8踘9>Ӹ(v4x4F!xyPNM5D+Gsl$PD\ɜOkW4sQu.oFׁ 1Mm{u?W$Q‘}=mZ_^ZT"&}FOa9sksBtj}ԧEx63⯵͝_?\', sb.^fl9vPVv=rߜj6XF貦%- uͰ͈&f-.A-p(|P?Oz4ow9Vi'׵R3EfqܰD`G(R-1-C9?j*ߛv{WȎyӟO?~9᧷7sp 8zבbg~Mݱilo47be &6ݻGڞ}coԜN)֧?ikVˠ+*aJєh25F|RvTPb; Z V4W+^OW4(PGtr:3Z5"K c̥4(-:NQzjTV=O,R ARC:]3JqAD,? _{&z3A&B-Oٳ 6Ԝh08S@ݷ SBC"%!B,62'r)C@c:C#&x=i!l{:fZwڥ'Y<}wy3܊.ABqsbmw듩X~uSE5ueDJ~rٹތNy{b-lKm i2jfՃA}(5HZM`kf*jK-8̓Os4Q%;\J!rc̥ЌrZ uyL/ ;=3ezi-hyo)Q\9 Sbܪ 3$ 04<@zHwqG"v ӭz-?PyFucoRhHm$#$D& Z[Ah5ZciJ{n7!*3΂@ȅ&0$C*If[\TvVyxY:.lhv`׋Χ9Ah)j;=zEg{&CP\5P^uA51G]E) !SΕ9XGb=vc=hl4jj<`K!z7XEwGgY#1hT3^rA3咄F ٘+Qٺ9r >:LS4)V^-?|p-Ϙ(Shy8i) 71)h H%h`SF .36q,C) NPQ,i2¢Z$|BD T h͙Bfg lBbR9KVPFX0B8|#2qvCcUãp[~×OF|:ӼD: L&jKLH`" ÌqS$U❠oD)~ydOb\IWNy&FKR"l2/?7tS6btDdp(c^JgK:UqþѷGoE#pDrAEEm$E31Qk:41NȄo(B 5ii҃&J1.$ 'U ՖbklӦ{Ŗ f}\ϭD 87?5c>(ݙ[0r.U~ b:'$Xs0O\%)S>)Suc87DB)#2E\ejwqT2N;qŕ0n,BrUSGU}WʅW/G\I++  We˄I͂[+P c D$o{}1G&jq-,;!cIQPij ә`%FLgr<1b/QL+ʵqߔ**',k@5˚IA# \h'azhFgԬR %+ӝ~ :M&b뒄E.>}1]]o)LF6?>T;-()CP)\p ΁܉M'bSѠ44fRJ# YPg u| 0o3) #--#@餃21ɜhv'K٥|»,ec6XvS6$n3ǝuXg69>Pư V LCR`E'f&%4SؙJӝ3D[NH\i\WHRcDj{qA-Ft+!e0qʃ`{4](s,O;&U]+;eě5G9uoh!d*}9^giaxCRXP $NXFE M,5p+vYXP$trjYDȳP~3r>A6:mЃb!( ,%>UZ P(k J`/i@yH9JIt!nKz΅̂:\B?ݹ;Mdq]#0gh?p?ю-^y$iBǡ TcQ}48^R12OIfAKSGglfPP`omz*8q:E ުy^_),)I\n|4뻟 nAuG۬˙|%=1Bλ. 9Bλ. 9Bλ 9A.lKۥRv)`] .lKۥExĈ2ώgPyM80jڣ-+! чL y\k-[*oCu|B V1Qr};T$Kɖ#v3q#vߧ5bl+ԆIUS rSVJF(I0{AFwF(vml1Ѐf!hʎ&e3Z&Q$";yN* b38UcD"s%mBaJ{N^V?..fCJevHڞE:;q6Z@%e8RXwBNy:C[CmGҸ.)LKEՀ.nxGC&l^gX$wT>YjPr& eca38Mc<ԇ 챼]OmY2_#F6_&sfC.(I;$ ُg~[p.2xs2.MdȏqeK>ini0p%s! IPט!o1%+0u,h.eLϴ<ӳpa}tq'J֖''7.:"|N'o@a-]$:tA8T3#үyrd;ɑf%K#` k¤)>zd] A0L4 'GZZdph|J%얡C rhZ;uV|%B~'bjWw||nqvO x.lʎ\m>bo|Hb׮ܬxxw79]t~\'|ϪEڬjK75kNS?P, sm[nGߜ'[x/Z0+~\*d8,ʨ ;Al>lU`~My܇x3;~2;l\op=e+v۷%+/8Mw G [€3r|)-JD@chJn(}&WHS7{)]G P6"E=]UFB6K-@C4RV!QR{WPm{r$aGpք.p*+`|?pHN(BQI:) -YdDIFǤC`!v%܌ז7q} ?GwG;5+'_dzX>2Lkuq69-^ޯ(dɹE&rLg㇇տ|O JILL_nf_ç8ScvMr/Ufu75Ű}bC><ٍE2mZ@~pӿV^%{MRU~ꢷxsw~3+37ƣēǙ}6֤ŽJ8:Qjot06+Ռۿg5]4Z$c ~v~%ì=ݩף&HuҘKbtb;08Ot:u_U:N 7fEXF }-H^gRݷ%{=4"-u#ܠNPMq0&~|ELj<3_#c!֝lEzd_޹V:d=httFAy tTi&U€H&pʈhsNP`y=%w(áA/< B0bN(r.uJIP-rųkzUcybҍʺ-/x;3GVyY`;^ i"<{mas4cZnQvbIuޒDˮ%H ]/tepi.cG*8q, ?${LR :ĠJr-Hr @[w(SL1c _ ~'.VH vr؀T,lJ(P։3£+Ҙڏm5;`]Sfݒ}cMM`?u2g]S]2 +bgͬEFK1H@d,{%-;8;xHe8Z'f˸ry=5G_J3ޣڭw]vt&vN8;{Uv~L=7 (QrH^.7F4 H(SޙDQdH Nxm"daD29٥OI6(97m*c@ZYIcBb*lх}V12 N:L9 }z`S8e'?K8 z}ϙ!X5Yk}3Mu$}[ՎRs!tuAo.wi *.oE~^Kqѱ}+<=lk"櫟mx)9Ƈ|55 ]}W3.yEHy&Q?|W/I*߆j_Q{:w~cu=(&3VkmgoܶPIb;031-'ڧ~dm3UZSj&SǾ)t(v5mpVǡKڑG5;ax]Ν݃BJL~.D!P2*m1AQ(*F%m[GQ(y " \=K>Yd@j=F,5PDWgg<'_L痡tYRlWs?]YrVD`r?{۸_6 ~? Ygr3$d˒"3jz٢,K-ibUUB (4GPŀAހq7 /k(T=qp)h@&m+8:w93w`]AIا7qY#*a\>ln8*9|{ɛT)>DTp=FTON/zկ8FR8a'jbOKtx7*"L@.J@YYe^Y .3[vtZykrwpSb#\[%ueg^sj O5ɳin`.FsfT*~SneU_7t鋚^Fegr3/| ?DY 8̠etV{8nna5:?.ZsWtb G܁3.b1q0j +۫h'ݫX?:.w)=А2&`U`>E;ḢCEYZ$`)` M:?t{~w>nx-K=X|mQJ<1L F1J"'ۤx&,!XIE#7&B.A™}j!|k.('Ed/*0$7v[R$ ȸieztuE?_yUQ>ew/#c@#t&r}צ@0! H 0*&WI4'„y]+o]]&kQ8*AG>dӨM%h6jG9ϏñK]_1Kx'g=K*S[ljoWd;T PvTk_Bb;E/6_5^EcM<yRCR'p{-kyԔ8ˬ`%E {fCF (S8ˀ;C=SK>z! ٖ;,ot,HCDfVkG:X13 a͉eN+;E܌vdƃϏu6K/c[1g#v:0=\֫ $|H!`N&2+\&j%DB._c%EB v@pkr8pȥP*QɾURtpaĸ: Jf \%r19JpWW3̟Al  \}D%!\FRy!< f˹3kVlWWfFX)@!l*QKvp JgܤL-gD?_ Rll˟?Ou:cJFNs2ڊeifۏGr-wsr? dPy0ŰUBSտJ@s-cz_+qB:S<`s-:2f芴|-y߱0趵!36):#BKVQR11n3@cĜE NVWl:tzS,e-Mߗ `OC} lч/QP-9βaK{hzջcp5~4UI,}{aQn4k=OS3qHLTdɟUNY&nwGx^HTu_ja9O5 -vQaƹs 阃QsgR`T=Qe;2{WRQuHRurHk'b/c* e8[-1<zLDe&#QHYu2 #X L"^ˈA&Z hnwwkl`-dQfu0{&0.iX  \%R{If6b0(fŖbת&)M Q7/ qfAH$*Ű[ 5rp6k1a[XY}IlaQ+i#u[$LƷe@w^_uxrRnA;nzJ氘V\wQM%ă?@*Ч|Z{9'\T0ahcQFB}%PG @Op).}Ԗx93l[#gjҭ*i3BB:,|R,Qmr4$UK7MEW r{+GlE,2`xZʥa:D=I4H8elTGՠELM !{D%$ؤB:`1i/#vHHR- 9ZFypL̽`}Amܡvx"fޥZ # qL!AH,خ"$vVgٮYZ!f`E;$` f8O#ȁQ1-ak$`Dl?ED2";D\khWd[:`Gɐdyaы51d:o 6QzHL8cu4jH\ pk[#aDrU["k|qɶZEb8^jpUXXcX@b@؎x U)q A#86v/xwleM:y0ǭf0GUV)﬑3ae%;R(ByŠ)qǥsܣx 1zkg]eUxD%PVG&)BN^ȀoDCsh]>jWw9(tfYv-sh9mÝ !6yz~k^]*n{8A3@xweƔs64 k>ӲN&uel [i|frTdu6g^Ӈ=a>hA5M'2؛5LS{eg!h4zl#7Y9S{YmDhʭa8),EmVCz{'cDحX\wl7 fK /NMh| TQ_g1-*HC5sjazh4b[:$YPƕ#HA#(MrɕВ2W$CF(4 >4gqf?_|5v8夎 MtR*5#I4ׅJE"G&*B~9s㵫-B] yg/^]@SW󖆻*Yŗu{q󢂰*H`/ͪhqL9Nʟq '-Or;;x?J89xl$fCk+fJ7lyp' 6R9;CԋR2I8JVqI(\ęR[W%<ۘe' cw-xmoxA&gxm [ alnKQץU=ʯeI5ڪ* w~wPW@Zբ?ܣVB{I[zغSQ|AoOkuM֐ж" IĹ2O{c!2K\;w[+NBa"g "IRVCNGgR2Q)c=eXёm٫FZS 1PE Θ '-9Wdi>a5N]w6<uB~2> }E b yjsݻd٫O},;q|e9 ;Mgvj%鄚aՋZ,%\k)m˭A*c־6(l% te odX g^ŸNz4jg66۸goy3hi$1mDtv!|_t$-8"+DžDH"7& eQ9,"<ϯMw3HK+;>=Ic\mA04YK$xlϞT=x|[0QIP&s!26"oڔ@@5yN鷎ףTS`zz5U>w:#7. 2LW_Cn}Mw>I4ncXQ)-47ȑzZ:%7;vMGN$H'ODrA ҁ)1nemI"48*hf" ^7<@~98~e>޶^IE#4_ ؇ŽD ?wDia4$IhQ2oTHe!*lV[@*oVV d Km`u4c&9$z(%),^rJ<5kpm< q3\<-IRu{h)?bNQգ^+ffI)jg;8+_Z^o5G u<"hB '+LTP ֑ذdž=6@٣Qsx%.<gY"ь(F#yQK$ZbB?׊"5}kNP~⴮0b brBYa"W^FL#I'y}XfUN%E&k\ˠYt4a>,ϓcHUuNjX:q_6XjXWB-$\Z%1gpt|[^F8iJF &%0DŽ)&ep[Nj-ɑOAۧK$~axK0;}6f_ #JKŽ&KcIUpi^hv& LBFF;e&Ζ|^TszS  +h$P*nj& MF̡'b;I,w-=iK.dITˤ4#)Ғs+# TFCHS+xF (DqU+P(I!F RѺa8əz"ک;u3+cܳ N/O=zyǏ^'R)#" ZGUH2\(u i!^dIP`.FguB\h&e1RHwxFQȄQe]P`Hu6qtއ )kg<4P MQJZ"%gF gf(~0M; N'3V "}$V(|o>tEYʩ-C[yb3ٽ"ohyo&g  2 lGKkZ_-m7zt ٵ;px'(Z Qr^h` Qcs'S['4qؙ>X}k{8R Rɐ;$ .1ƙ3HdV #lczkPA )R' xEB7ZeiQr5@d5X}@1vc}d[Yי4q,B2. @J+F a /j!h^AKw%CP&ɀ:Iada%EF+EȆDq^39|ԬH˲fZh 5 q-?IDru'FǷ"A89e.5'bQjKg)O3fNTܨ )\Pq}-Cb49?囲2 }}0?'].+)RZ$'uLf ǵ)Hsࠣ2krQQO`\t0D#xu;_%pf%ol$F"DX\ȭ/Njśֵ- xpd_O5~^i^:^]~T&T&'_5 Tkyk%.!q ʙCkO[͈,uUWޕތgaP[av7u/vlgE8պ?K۝n'v_#tYҖ@,njY3li3PEA/lfvGE +m`ż@nl_+:/mխN:e}I2muGĉN>?z4@Z#~FOiWgCYPsanuח8/ޞmſś/(u~\u~F/"絛3wæI\:M:{=uuvyIG[- mOR8yM7T{罫뜈ɛN&v]\wC+Lb_UAˏQJ;RI;RPi@`_mލ[{18c%Gڤ*Q/=&KEPGTr: gށ3Y"K 1Z2L"mҮ=W;@6<#xLgCrxKO9>۳ F0N-;8]o8}|~91l kY{;<Ylj4iW51LaB<ʬڕuʊ c.(\ԶpsET&&̎*֖smʃKq:@RLDBHj00:'zU/NZ=~ʘi"瑦ۮ(7S W'hz ʯWiOW\ttO^O/] (&kDp-/g0E#Bp_p$O4h`g\HmYn!٪q@]tDtI̩ĥTL$А fgBL lBab% &mbLrw,x}죩]mtK05ηE?W4]+p. 06,bD$H^Ds1ßgV2hlgBhƽ)Ii?{1Cs,gm͞ Z4Eѿ VyL^ BdpS@mha7HCleCeT;\~pr9RU.@rAFE=#I\i1R*NMBDŽ(d TkՎF o4HOW{MzbK'E)-N-hjY3˂K]9tol.h31<cl1#]| EyA(y_v#m6sxٔI]Sw{+yߞ Q2W$u ]|ّO\7<ξr6zm aFƀ+FփʈZ8eggM\O6md\4gY&مa|5}2fuORbb\.=Qr|mR[~. BDQU/ϣ_ZyJuZN1 t/y? sкݹn5jMkm~InwI@-6eѸxzSfHY ?es{U?n!=6%ذ&h꜒]%].4aHU&xe5?Tw *^gq:̒W9"n2Lӕ5W5KmHkM!v3F|%}~X>3)t~ tx6}H)3DZ8;{"b|~5~wӼb ]Z\uͽvcZ<9Llֳ4,P؜^ΠoS{MI?.;ݚWVDbŦ^eSm4F\aj4- D2UJsF[^:1G)D>O:1`~H B؞r!kd 5` u¼`*2 ~]:az ]<ϋ0U;Í[&nĽt'.rR&={eʅGϫ9&Zo^SRl]{R!TL:\r lR'pIG_,A)H:BMP߂䓡\21`CǮ]EU-v9sDFhJhH0I9$ [V@vNi D*keȞ5>h!N5;Lʥӫ-1*&1u=b4+$ų(u{Q{L,c05%a4UTsc%bD\5+; zSphڢRa(HJPjD$HBF(rAcmEa䐤X€/e;℠aJ^ '^q<;~k18/b{Y8A2miK#5?nWNs[JWG[e'RW*QDF U(9L#'hXl֨ni (WK b O]c*h-µЂNȳi3pZjwEc)E;ϵR,8/XI LS?̥HB :,KՍ (M|CჅ\ f61Vq$\@KeV8e*"˞3)d4MKSַqA3]7#ןFDcBa-]rW**oAzđx‚ RϞƎƦ%^{ Z{^ h&i+zsP0@3$ ##GDc@ 1)/Q%D |5}j;cEp0gylEA`>:Ḣ E.R cS˽'3y,Bްӝag8V ߇bd>%ADbLTJ"'$x&,!X ,᠎!dTxGDJNc4AdaEwZɤ& B' LGXفO&ԫ+vW_ɉ޼9]NNW?w22ʜV`y&XJξ4qg!azJacvkNuǰgizcp{ ̥2Z[kb(T|5+r_bqaD`Gb|HMÐa"ì" DFG Vu7.c.-c6JQli\8i$ }~O]d?'8WwXT- ߪq @?Ӈwg?&>N? ̂ g n=4]=CS ͻumU]NaܻhKvgkI@~~n<_wyS[Еg~JVP]Pag % Q C&tVƢmMpRQ/' x^cdpZ5%βwmm:WxI _/SS55km;k%c"ӖwvmQ$t7 IFI]TtTkaP@!T;J/qļi>?1Ae)D2Z:[M^thb ;Cv#Ȱ1eGkvٶ);GUϋ*fΫ0svC|aVgm@Q+b$2OOrRX[6q[_^ޓlNz.4\R+ytٰ@(jۈr1ĘWXTmu5 jl R Su!Jܘd8 Iy% &w׻n[_XSd=N/-W,_j&6_ݺqfG;bGDdz>|NJT& Fu<&kPfJJ(Dl<%ʦ$`8g[ %kR+5% %S {\$C)BU{˪t_ؾt&u:DۖGK;bAO36óQ>=[l*7t`Mߊƾ#B0H3$e@ʅ$6^Ahf?< ջټ5.g]!(>[n,ۖ||W~թxQVz|tiV8:n+\ϟͿ?|=ްm;Rd/W?-˱r?*sbMrMN3bܫNnCis߭E6`nsj3 y;K6_ v屌 .%)*b?S˔am6F0=R"!ErDC  RA-M3$#j˽|sCorb #~_p9Uv?2gyPq׶p_ [8{P]Pl2 6mA 9@YoSb-F_,-UJǭn SlhUQ''gmӶc_sлS`ws"lG0b-9^{(on}u) +kl*i I="ɕ6gksCf c$BlQ }DaNH=1 9{]'13_ _EoY'z_)4Mbez[>lַ vY#SIK] g6[mu`N1tb%} bC b[!IRէ[F1v`f?>];/.nz.3LvDAEz_S,".|bN]HQN97HFswf"* xb7).'__~]2Q>Dڂ/p }&-f H$Q(DZڳ 6`/g %&PȄPb˕zrk39hQ܏丬{`G}Am;`@1D.NFmdkFi.iJg$' y 1'5BVfhQ 5(֔j21#)q'4_L֠8z;0sUPG -"`Dt3"Έxg-Y}lf&N8B @51+hJ[? qV %kgRIR%r_ `1ӣyFwUuR\<.krFɶqf\qD8sΧ[J1b5 ,)#B>#Xj[Im}ac[<g+ {oտVsX?dUŹEps㎪(:{7(ws_TI ZVYQ`>zu;̻n(sgsY _p%+jGo;np0W\6W}o}n^!\9 (J V: \ukwZY)3\B8 :Z+v_#\ap?0Á+5Wաtϝi{:\`5•= Rك+5ZpVfzp%Zź[W\ iUsϣn6ίUa#FԢws{d‘7?#޽uc^K*ǧrs'/u9o1S&iz&&SgwzONs 6"QI912Ԓ8agYb{| ,6Bٙl,5$@@}i9?+BO7k}O7kNjtiYʆuXN>j5V}.gq>^#@ ԷNMiJQTՔtz8~w=]"+||Z' }^j(uew}3am`9 uB.Z.Q ߌ<&a՟qf;MKˁtpiE~xfHo8,∠4ql`S0Z5ĒQ1:T&m+bhma<0쓧W) X{((h %Ea;=7p&ɣSlꗯP}aM U^Ƴeo(W}- 5NME/ ;-F kEjN/+Ny6'#rNB@Rl$DժUӁ{8I:nX(c{c˚.vkjMlHcIƏ{2 xAɿ7mJ'&^)E2ɿa=ko=E훣]z\VħU$Ay,VέL3(PQV ^ .G |͉qq>ߗw_,N޾X}bBX\*51t*杔I@AJ΁K%l1p^1ԛ4]NƖ1h$9uؔa,zXmwHn@=@c{];Cm`ZYJq9Ct@b2!ppV( ױ3^'yd_Խ4*%X6M0Fn䣐`4S H/dc:NGGr),3;XGڦDfBS ٖ=bE* Thcﭛi7 J8j*~3>[ԣtE~2.n%Fr'<_z_ψgG'_Wk˪HrvK1'ug8s?{ٻ޶r$W|/sm>d1@Ȧ{6n4:=`g,ْYC#Sqb],sŪyq2ڞ2؊-M$dؿ7p$2PS$4L/RoٳYr@3+ %T`aWKp4~TfQL{7H[Kt;s"oZMp%$I|UW4ppѵQ:]Avy߻֧Y޴x13ү铷oWœ/'?p<[ivPDSmŤu}s+.+ZZ[:]ՌX 0bLah(|8_z|dM`V[꼓Z]WefV8HiXpJc_W,S[zTvM? {옦_ۣ?;監wo_ӮI 1ƒEaijo޴DOӪfg~vfW{{Pˋv[B{{k|Ad<=gjOʡծI_!i.¬YT!&s%U8 AUviܷdײHķ;%JG 'T3=m^e+E-'P8{%=Xm1wD-s״>Zw7@N٤1H,0ax6|v|.T9ޞJxvlor֚[nmr}x5]uztWu>LOÄs-K&>+}.YIZmp㳒IgՖɥb\.{ȜQ[bCԇI)S 7!6L5>k|1tJFrŔtH[>zx~NKCF# D`hK(sWǃKIDb |Dlb,}*a۩ebQ32usCyi]5q2f0W{X lI#@Z>~\8cGWϧZlִ[MrsӐ9@=::93n\ƅ<$F9V7Q6X i2:0DD?$!ʞՇD%d sOL{V%O.Mg1Z$N ]~S $[^bX٥o43&!HARk0֘Ř? {AR;>}_؆ >I%aͬRzB@ZcC@1 ᮢoPDZ|k LX!Pv~=;M}r.]7_ǩwu@[q?:ާ^X| +gbjzߦAn }W?}7| ?VDv68GWnt+7Zr: kifkvH뉐ԏzr1fzxek, vץ8\Zm 2Q{!ty]4]}M#HH)^*PaYH#Xcˍ&7Fۓ)O,N,tvóɧvv~b.&2=ﱍjs&nYcfTL6cP&'5{np;t<6i4uU&2NsmmI a*P6{"z$\HEsg g}x4NkCgv}uY@ץӹSr 1Ki3;_{)w~|}=xe My"HN6^ h΄|;ַ3kyd V dHS%i@ (t2Oz>`@:&ey,G^@t7MT=ӠN |Enc㺱U=oR)-,0fn5 F,a!$ԙnVY3XBQ'o2F$y9"hcv‡zGK sԊEfv G#O!IE\D!3VTW\ 6$Q/+RIDګ j%慚> (KӨXG/o6O Jj+n2lӴ/B[ȇ x*g6Tp<+in?.jNy 0Y 3-6P&H$tRYl22dYףV1fYxkĘ%zR(&fEt)bLk8W2 X[3VgfU¨53ԅjB:]S]mrQ 9rfVi?wd3GKeؔ+f5Ld]@瓕6%upu9@PȞIK RRk dɮğY4V2 ks%f퇃D1֚y;n --PgcHT#g# S9fwթXl~7ulVڸQq LȐ !] Z$⌚c*91d\٬[R?#Qhjc[(+kDiN#en}hmˈ4(<.䨓R&Xu3 7N $Y.F&,'Npt-qq8ѫlq퓭,V/Be(;u3X}k$%פ"#(]I@yqK2>wzqzq_aٱ>T!lwWܽOiGn>cuy?>#OKc<zL3>d1<Ƣ!Fc]:r΁[ɷS@,=U+IŶt8]j6%#Rƽ  ym#Ss!uA*K Cң :R#2khXe6&=t|9Hr\hËQX-ux3[D{{HMϪCURL)uTHH`R8Krv1&xH3Ɬ*ȃ5ajO{=,s4heYj,iÌ#bPm[Mo] nƗ_E%Ҿ։ i^ VcvQGI!s@h!nbU-pmhw~{Y`e󐭊!eg,SKeY֦U_.o.6}ny3=aXWhag1_Vz}֍]dt:]>L`v.oßN/.Fa"{&Uv$ARZ3Mh]i&IJϧ@{XM8 [kW/^Nm{/:3umoC\]+tVymuvA{:ȋa< =IBص 3mwկߦ~ni/gO~oO@ _ioz@Bߙ\}I KobJz^ĞÃ%A؅%:QoNjo +A@f2: 6N 7Y$рdFAEə(u xYӹu?(Y:ebcmucu)o,􊭴cOfǁxXBdmYsFdDu`&el>cq@S06.pt aP!1j L]~j6 }N7"zu(9f7T=ۅ_u=w+!׊6.9 ړ RG7V5htCʹi8 .ʬVH%*9zuj ﭚ o|fGK%χzu}|9ז _ϒxK^sϋ׏`E>Ihyc<]'M2?7Z `i5/2?Cj$`"+h xSS SԎ3SÐr`y#X~ [&6JsdɚgÑ,GT.0Ɣdi4A8 ˨5H )K>&e4VЙ IP&йleD5qiye)S4-o4&޹?7n#y_aM]en6o*UwIR{l%En~ô-٤LY)7-m~1i62e_Oo=V4Ϥ|{:j=87z+ױZ?DIW{ hFZ w*jU`_+VC=36M]Ot6+ث{ 3: A} ~^;bS{QҚۏ\=}ŢEC$a|4~Jf{dJ֖q7j ZB38TՁ?9xb-ۇħӢ8[D;Q+nBO7ԯ.cG8/:6꿦7ВHxW/}%[Ԏ,>N2'fՑ%g_#tcvMGu*jͩk[^鳎_dcMmD1bn͓˃|SwXKD YdrHZ2ˌP,3)5OK(dŰcBE W=Om6Y&{6)Kv1M<<u.,7Ƒo0Z>g# Lx/ofIZזmxbsuS'6|dE~aY=Ϝd`Tdh=i!}҂rݼb<IZ4vaK l Lv"j*h$vMg״5-;D;僴X tҼBe) ]Pd{QV F ft`1jV%RT.~їs'j}wY_iJ,u<g"ҸBtټ fELYXTi==P$:kiUz *o[Q{ͬ|VQU53wzFlQ: }o7lrrO~"&4U~w߭=dEUwa^ڮ \mt,/m{Xl;o]ѻOס+V+d?+ٞ1-y?g~N3F \y1-.PGu: ^[nZ=e?yV~ exqEC'g:kc|1!wi Nu_pxg?˟fjjY9onܩ\?+ݭmӭ )%Y0LpK- ?mm45rfcNs\:uhk\X)9.X%5. 7qow3DMjHkCS@)v%km<>5bk;J//K}+ @\`HcG^<8T!P ,m0$X`pEru0"wWW}ĕuɸrCP?A rkS^t&ѷ3 V5-^;SJ\!OK3l1vw0v!HBo\o,BNi"Lڣo9I{i}K.ۣdOVgPȣߢ$DVLU^]b}|T]A j|eCQl}ƀG'ᆴNn/~lTz[ޚZl/zzhsḍ;ѳ/_d6k{߽m̴˷tܼmw0,2OĻ$&_31].@U1wT=P`swjzkFҚ 6ΚʵΗHӮB33Li#,c! s$PjG PRB >JP.$\`К=1LԚ{WoD+'% j2H-W(0HcxͮgL-O`OvD8Fj)KIpLx\Wv=( +k \Zg+TY0?J  *j HU]tnUq%B+, \\CUHe\ ` +t("VwW !pE5 W$װPpEj?$nX%:+'53Р_o j %Ez1Lf?V]`&\\Bլ"f+a!yW(Xp&$WPpEj:HRz+阓2 \IX0"F+T+d+T)pC\)NLwrK~b$ a/'B+mQ{t5};5^1fdI5lNfrqL;QF*䀫CFi{BҊ`prի""\Z麎+RՀ⊃w4{-Ԟf* !hd@":`&uWR01ઇ`_wKqEU8kW$WPpEjM'ұW=ĕ\+l]0BPpEjE+Ri流r+; )A sN(w_4zӉ栻Y *]OdҰb&3!e\e^0Ŝ&|;MlS2f@ Z2IҨgC+B !'?B\CYZ +R𞹏K  * \`iUZy\J3WV0Bzq `I Ri\WN\0$ WTJ;JzuƄa 'n&ԓfj͉pLXWjա]tFW$W+RkxqE*kիKɂ r H}6 EA8ίѨcη˫Ȳʿ/+mB?WU]Rvt_dU"l`?+܏r:Gc^hpZKn.DpɤR^~˞w;a]V+ƻ:ѐknoxxzm/^[mt#ewMo7=U~5__cSu/%ʶ3(q Stz,{a7 ph9qӼ'3z@'>|BUFN{Z~vQ_qߞOY|+X\K[ -rVd^AЉZ$DhH*SBGJb/˵$I@<-~6ϳ+$?Qy`s2& |Ƅf\qe|d</^, ˴Ow-q$IW }))'ۇ}/] M"i-0ElwJ+.$>",3-kgd<ҰwJc;vdǖ{J(D+T 4zØ aB"ZlD(cB]6fcta1 0iZ?-CS=rx9$R-BU4$n6 %F"RLfIG>>>ǀ͌eEjƜEܜBёF̳͌RLL=E# |-g5/ \1v03$^yT㊳N,`ΐ?!ŜkXhK3xB)b14w&)味8 A>14_Hگq欴C69ҬkcE@\Q,ɵk%nqx(#>o.BMfIʘI⬉@+iv1&Z-u,5 [8SlJZ{mߡ̵s`Af3e -GqÚ`M DL/ *}8YkOMEO6Zj_l5TmCrf8L*|vSf vn)YSMUdb` &`ՙN >S.ډ`A֙/+~ E%9PdS REo1x&X,*>>N|kBE?GX.Y_lBuZ .S cʫmc/(mik-ӈGn W< ܩf Miݚ膽Z;{CE hDet((*it2km ,9 vb[C{աR- 5 #̆ 6aE617`AVࣜ-՞ꇱN_4*Tf473A^s`$-1#|+q @f3LH HL)Xsv.*mT -AhWߊ1;} E@5!r - DejPVQ(Zfσod̰B"rP:׈7rݧe6*I@B4#Jkl\ u"7bpntg?UeXNS cE9aQg4h!0c;Ѹ{jj-[4=>Ug9D; j3kETJӴANu3{"`!-V@ax/i>{+]:91AZ@2؄u1~;~>/giګӋqǹQI r`!a'[-C ]4B{j@Ūca`ZKsrdns^,Z,  Ę +@99+ qYaSC^Ç^vts]^`7;t:0ɯ)Z7t*#X`[aH HqE =>#`\ߺ]Op6-:'XWHԇ-}ZOnE#G9z7*XV1 /T9Cآ QbFL#,wuk[`rM 1յNXG6# Sا}έ Xv=Y[n9l!1mx*J5s tkIvyfʳO!"'BAǬP /5W`20h|/fW'iqV0]p(zOYgp8\Rw$9ЭxzApsl6K3 Dl0rn;V@jRKp$2ub&H`_PGQ1*`j?.^%Bp>/>\[/.o.<\FX6ɚ=ec9wp;?xl$Jtib Av37W^o޿mrrܻrWr /ߎ! 8Y;v>P~n.XNP4|rqwѐz'k[9; 6v7]iXw/Ju/ozzyqkoN^.0kO䥼?K5\ޚ+y볫oqy:`W+al(x7w-ga׻åuLӥL;M'Ξ#+8ҖW \1p;+Go3Aa ̚>g4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJWzL*uoqtve=p%(IW2p䬘4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJWW \I[ \܃N pSL[6W@c*+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4pUWBʴUÿ1 \;'\%9 \=Uũ+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4p+ \iJW4pjW i!L0$?~<+zeԇøa-w_vM Yx30a@y/"q>/=>D﵄.2x ˛+VJ~G~?eJWϐ-opl-1l׾aJPrTU̥/ OAmxcWfahs|zJ>cWtt-qCti3t%p ]YN=] J!]9J] \[+M;] J9O!íBW6c+AVU0 +uf3J72zHWM7Gyy,^c1'O pٱ;1/^IS U6*n[iuwJϐ|8\컫r/|s!As|{_`nr]P!>LJ_]_k3g*lxSQ OOWק` >ބ|s67{3ٻחھ5rWod/]{`W N_3nzcu wzu Vڷ["Պ owafx_Ct'WȏZtO#**-g&f{=xkcnqy/bMwSVlMs o9C,Z]䯻EVVnNxpss/)7G\@$C:mb ༝.obm<̂ҩ~;[2nKt%vNܼ-=]5z9 6CW7cUuK^ˁܓՃGapհDWC*=՗.uW!õ pMDEW֦c+Ayl oBW!q;BW6l!]y]ޒ`wO{\BW01)]=C" ѕq3t%pP2;]cc3m yCt٤Еu~+t%hՕ N9St T`܉YIv-?)F!sNU%Pf#M'9l1+AkJPgIW[?]p1;˴]ko#r+"`fY4pX d7_Xdx/濧zX~9m{XVSCxx,NW"]Y) oV,1ZlYA)pϦs& gWW'_b8 8: >: bm]t=] U]+P "+tUJv*(vpOWBW\;CW]֘UA7HWBiCtE׊ J;oCtE]3sWmmIt*tU<ݙ*p ]tUP^]IE])_UE4?zXIͿ}kNp19\Co90ܢs`}K`k:s A-ݗԜfm6oݤtP+[q9@5he33S ->:[Z0%v)+PwftUPʞ"]a\u)6#Q ]Q|R>-ҕ5\q9KWX ]v Zl=]JuMoϙQtu`~+O0D[C-@Wc1-u`]Rv ZeNW ]qcӳiKs%9ow?k}7+y?{:Z?b>ǣG&tl -p  dOol\_ל7z4E_ց7ױE.:7+Oɟ"f'sfj?Ԏ3GowA(?RVR$Y3V29Qy]qC%~.?9Od9L_'T+'{Ʊ6ǿei[[~e膃?Sܔp7lMF~ W) .zV"}Z;On WI#%{W9[ 0Wٯ)L"k!2JH69u$lFr'Xk%٠+OEWqwe'} ibVZ|o+Ǖ#!FiRfb-hH%VL`%S g'ZiO }iycGnEdɲ}/gdbss~–YXaތF4ҿv9SuT>xfW|.vUeU)ㄨv+>ǭ\+rW42MM8曐\ ʥR(ՔS=hÿO1LRίHr{[U)ed81,XL,TdÜ 22 i#6y$5.ɾEi]y_OdVҶT?mѷ_GVpAxyP [6^}l+8TY%{\rYfqynXjj [tݛ8~qv(Q>0s.bzz"*bln5n^~ Ui'JЋ.i")SFkd:(QD}:gP2ҖⅵhD&V=M> Fneah0Z]JkI c> eg_lFƄXSYa3-鸬`l"Kș ^hShZbd :Y00c7g7c(QMYG[Xz6ZX70*bA%2cV=:Y;M 6+m\@L$yA$C\ '5 G(Uddkk~Gc[18EÌ{Fq&m 4!99^S2G"1Ԋiƴ'%ج h$υg|eI Jg@ttSphw=Zt6u6f%lEϋ=/nxc-+G9"#Q(%Z-'+ŗŶacq,PLJ/@ay6콹ҧz9؉eomYsoQA{я'~xc5HŦ ";,TRd,r X67?dp @ x͑ bGn=ncJ%5(xIy3F!SE$mxE} "ҥ(£ ;tPIGRex磱ei=WlڢDY; Mowͯao"}?/v9g(sۦgM@/vM#A΃f(ZJd5\st6RW'^IbeU@N%˥S.-Qç=vtj@`G&s$#9Q3"ȳVy04;?l;ydg?3M* [w0Ci/N$Kv'&3wpmW+h%N{wc_`*=|܀o6)RڲJΕg+ Ivo*m 2ifSai*K|aJ v1=&hL2biTAff;њq>׍*,bݑb?©An6-w+յ+~KqzN0_%xAE+J]橷gO9tf{pf!(rn'9'ޝY#絔a2oot}x=8:rGoI{Je}e/G>YycBNL%Zo^Xч<G0!@e,GP.0Ɣda@ 9Fg&^D *˔e jbR 6<3=g&&ޅz2zWJm: 'Ad׳eēAVo[)#~=?+/\^`g_L)?e .P"THdO"Plya=1pbx2c+-u}2 јOc6kAdr{I#j)CҤ"Dά-h:Zϕ9e2 xv\ CcpG3 @߹\1}32]xBn7gt[-qs* y);#=^sX:T8cѧ뉧OfW3Ϗ<@J9Y2:W|пrpyBJt=4l O15yؖ8$EE>ivWtuaRpAeDE-鬧=~΢ݺu: .o,;q>ކX a߸P/Um3\z MIs?42 cmP|woG]mEe9џ~ӭڂt^~H3rz &PFu ݶIl]mt6yѸQop[qݏz~PW^Uk:bHn=VY%xau;sJvv9?@!vCZ)xa5gNww4&MAn̯zM  J&cεRZCr|H!JVX,,ƽ{1f jSөL}ye7߷q\ַi`?z Dph_."s'UaY!P[7rN7@XIn"5 Ul^Π>ҧNIєTꮹ2gŃN?фEjVn*9crls{h%R[e+^!I93Jpd҈}xbBj躬+gpvfB=CN;!28, >HWhyy֟vW0)Z1cbSwNY,.-\+NDŽZt,T ƍN& 0l',KBZ;!k-;=kV%Wd\c j́G22>9;x?~lј;9":% .!DhƿF;f:mGZ8\VF2ԋ(R>j[R˽'3y,HIkhΠ1;M[8ocIkYьA\ΙTh;n&[K]w3INC[2;LjVI+ׅ+$`-WuQR9UƗG h,!y &'I"9klQS,d1"HfC)gB:l2VOzlKrFϝbF@ 9Dt`hv4)Y]7hChN,ku99s`lΜ]}l:y擞,êu8;g|ӗ2w?ȹ\.jX̵Q*g er IqfҶ#۝܎vh/Og|zRXWF g!c4qZfqR^ĬW逰 9^ wE%f| *0Nj-*9No9/vMV픤/5|l*MS}l&6 v m-:::MLԇEG(man.Y @"/aN(Esfͭx{bdn!GD?2$Yȡ`HY $,Q l烋ҥ* WJ$F07 (Pʍ"QYȰ, Z)Y t֪ B_A)x/ -9Q1y0;檀%? M.40 ݷ LrJXHĔcRt;<8SgtR0U"ߕiwW~F5l}0_nhmS]ILY8ݗuWy+e93Jb^|[%Aӯ ~YlVM1Y}ۉ\,lYJqx6 -ʳ:L>*l٤ɯU0}QӧY| pr^٫7?e Ihˠ]OY vcvfP z_Z6w \ | ^E hå6<7<蜱jJX)vx'Dv5푲}:ٍ>5&zΙ^!a1jg:@aoJ ^ߘZԫL*֬-^NvTTZes3]OlU;y=S~J(kiQ9aj ~I,%1egB:#l6ӘG"h)?eF D`A;A 0Ut4 )r6sJźٳ'_:djm`5:Ei[W8ƈ4F(noP >~n@{$DcK= ,g kcE[OT{T&(%4m+@ZGh)$3ǩ Z38+Z?̓1/s[1'lɕ}B#AL ތx6 ceeV2rM39eTxEBpyH^nǬ"PH.oˬݱ@ɯ\vU78tjnp V{i}HrzϠf2^+k)L@yM>jb w MX5u[M''.mSl`vYtmU#0J%(LqM.d9qwDЧ.Tu}\s,p.](|5[VΫ!٫ג܁DJf@k:B)$~ L l%7Kn#>Qj|bRwr1XR:p8ncF9S"Ist*lZa'f",zU䬑3ae%'1R3Pw\F}d&#QHYu2LwZ D佖豉hj ih71rnL_WX1ܖH-@A n@Vuz[[-:YV7MXp3AJn躒w>ajz5kIP&&]^߫\L>kΡu9:^=,g1͆~5e-jn+c=/\RpP}ӧsf]v Oc[fф¿35Ef[s:im7vz1 {ε5^ZH9gubǒwtMd`;K<əG!2<ɢ!o5h}?2d?O҃IAcQQF ;]moG+aG2`Nr.qb*qM _ IȡDc@vS]Um$c."H$N<oC@G917W{+S 4K1eIE綇MHZ;e A !*1ۓU()|BD 0 (9()+SD d٢ks˯=yE$xt[?Z;pl U}hy~;ǃGV<,Ŝ, D#ujwd,LEVRm!ZPA -]L^txFEUO5s<&I1P$\ ONrT%?!R4x#Q͓u뇙WroZ#QAkg$ӯ@P 5*\k"@H(JgD~n\'B1vCn3`ļ+~ߌas d+ڿ=}yeX霒H`? amچmg ayH ՚:fx>Y??rNsf[Dokl6h]eC(qLØ׫@\\dUƣ"gD<*|q<9"-d[tk2nv;BRkͩ-%5u7 շlU_R**SO6]ǻgmِ&qܦ\0])($U>Kgo]M1Lн80 3y;5$;i9`QPٍ&iVu3/iYjsZ +}fMfMR-CjJ/g xɘQG?(=泈Nt_E,9:i .g#)M bkie1$Ġ+r+LmXU}.xX6E/r =(l`yC$OTQ|/,D:βg.4ǝ]# 6i^N$!9th1RVwZ;Wgֽ׉kVJea}5Z?aMg c %ݦyё{RIV0CsO3Ru0z.Q83! D)J%b6kT}8݂M |H˗ch| N'핗Km,*$w2*yK=ZSzpBKr5-TnE)$T,IerqXJ$M2%K948-^Y+fohŤܩls0#p~~ya"u"C w=Y+$U~ˑ8-Fqr8Uf'T}2&o͟wC +u%;4D+EeB}$}&h qB !(L䊜k#L~:LQ ^!k" |Vt6P<&'hrCn 2gubu~(VӺUȆ [Tc:}?WV$U•}5m6^^<^^DL2. gk){4%r'ţh+.Єٍ+Usg]4Z~i/|}=}sKǠJEnVY:+;+t]ZfjI3]MZNsYޢ Xj+4yaT ?zr3r0pY57beO=[m5]>r˼wG/lk-@+~*|34zף Ehh9t5U߭Pٹ_Z!vs:0xkUb$vP$QF:Cdμ($kEo/.ƘkP:#qc*c :yH1X#x gCr'h)\ G'Y& N':Gս38Ԛ}vwGC$5nU4;@TӸNݙ3>ˋyddX3xKq½0>/=EMfNՋܐN3r4FE[W'GmGOiݲJx(5e g)\Uk99WW纼x,-?.A 7քxmn]Ni1GxE^,1qznizY|TUmf+?|_-4%NP)njav ݛOO WчA ί^ti8!s@:TYg\ǰ3r8ByE!Qb [Gnt{q=uQ M8|Linw.{c̡fAQW>q]a%/:pVxP%3W:p%W8KK#D>KdݻDD4Bzr̀.Dǻ HMXDJ٩Kĥsi=IR $J")2⩣`9?L+Mx]gjc:bG+xA Zo{\]>a`+\b&yľ])9= T殡6B>A>iuwwh5yBLEq]Նi/]Hok߹; dS๶Qy{Gzha\-_T_/aupm@(>^Ī}L~Db~RE ~wuú7jjٙϴ{Cl==0^am/)[>;>f[I{źehbMwmi=0ٹuiJɬsMi`e.q[u%XYZDlJKQN+TV Nwf07(ΏAL$PNI2A Sbܪ FA aihz[OxyhSr!K8tF2T*a]oP?zd QDQ CHs%H%:{أ=)z46N5Q XR(F,ku$0Q%39:@ ].igõ!uRzv??| ٺø|.RIC @{u紋 !(t7ٱJ. E߬xLysHEX6@g.Ocs N:6䕐c+nꌗ68-G\W^F|8DžTh#a"S1)# i&b{gl{ȑf|q,vmٖc;V[mufjguX]c?LƇG˻KEO談 مّ c:xbKq ,jEIHBШOxyraYxm]5pI~I=P8l| _-o5U>{^MҾߍ+\YYc _ i7|L"qd ukα|U`v0:jatX[7y^ wv_7r0b`.U9a\٫.(;BZ)Ns|Üu QQNx&VC5Ȉ-.t('@"0 0X7r(X,Ĕ"'$dbF{ 4,DZcFGaL<_~%]YOcC/ً?F^Ac.$yls;7[Gm0&a&ajd :(I%ZjiDcc,SBB(#x+ؔfTs6H'\p@9< ,d)ڪ\K񤃱s1[[0VȞR%?!jNv/x8!#TlbSt!,JBJBAО'nfFaI 2 πB1T=1cIbS!>eqbcyzw}+6KyO}HzmG7o?TfUɝZ&J&OZ%8˕KD52Lę_ܰ\mjߠx8yYrXɚ#Z "qja,o5*b,dQʙBtM6`UvIhQ8vlY]jrMisQ\.\T3g.~1-.dC cAa$h#LB ŗTx8x.<vᑀ/.j`[.=xDy 9qGяJK˜X)Ҫ;CtBMݻ޼!Rea+R y*{R0ϭr`=xXNiR 5M|6 r9*'LR d(a6R)wҊ,Zuw-P!&~yq͆S>ɲ҂%+{Bʽ1vX4_}uUxt#] aqja)¸71ݝϼ \S߮>OYwwcZ7CzT*OX]gSoG!P lvO=DJ^lGr#y$,p֨\4ڵ+VO{*W 6z`3R@}Ǹ(nq5PjZ\ T91\WWqUK0ʕ ZSUUi=ĕ%- B5+VqUU:qBA|oWX7- V [UU+i긪*f\x.IUL X ZuUU3jqEZٿ}Aݫݧ߯nAW-< H#'??/ΗWoM|5OsVgz 4}IյO?!oŸ&b'@VN!(/yq[B_ϡU,jX3OFۍ^A/,sX׃+ذPlR;߁:Pٵu5Lq0WT``sUF7+U \U/=J%f\!$!\UX.f&o]Uf\!V+p\\)\+b &o]UNϸC\2ҩpłગKપ}I㪪43ҊZZǕmWajw0H%Xq"Mn#PG(c +@ݼЈ5".rCl??,7kE; m"ދr$aw`ɓ62c0Qډ7 f{p*Mۃ/}QIߗ˳-KnX)/΁}qr]xv>ݎoT*/xL$rx ۃ{J\HW<в! +)X.B3/VK3u BG qء*XfpUb3MUk&RUq?+Rv*רVpUZ:JW5'RC5+ ZSUU)͌WrîB+pw 8P.8qZ#\ S9br䌫v=hC fpUJ ZTSUUg\#4$C;i۱XBjWq*3W!\UԎur 4 Vj*g\!fkWUગیujQȩ㪪5jq*@3riWU-MWRٺK\:)Pga$ۀOfҐzj,nb3:&fjNOf٠IkkVkf٠Nn&k ]9K lhiYZrm3+־tJ.-X ҴZ5+vzZSUU(}ĕI^lCiW,W Z9=U%>⊬hUL]v*VyqRv:&!v~\㰘ajwuJ;1J qܮ TCkWUપu4u\W+_axbZc3r-q*H̸z\)gd])gAB3bR6+U#qqBG`!\`eu3bH\Z0u\Uf\!4)A!\`PXWU.6c]U~ʡ@[?#쑦CP9x|6|2*ش*G U\UDt1mIlCܐ\f8XUU 8jqEBɆpł5fpr VpE:HzW+ܰPe'Pe`v1LuSkvd] Si'VO  g\=AI!\AMrME 㪪rJ D1H0H \ՌuUj3u\Uf\!)`'UK H<_TqB@R!\!h!q\)ZUUUUaJEXyx0(]UR~⪏csE>,O x,$~?| {y/JJ?Xzuat{`͏?qtć^W&"Nm_ R_+I(&VpBzz9/O'.➰]h ^C|'V ;7iEdnw.~BD;|Jl;>kcML;r嗫yYo9>?9ހ~$;ա+ FKހm@Z6v֮ S31Rւ(ީffjo;@iѠN43vAݹ^~Ή-Գjmcm)Ȇ;l'5uժvW)3$1?uU+ꖂlX \emȮ0SSKb,aa"OQ(ۓ=JL˽ۺ#`$MTխsϭӽbQi7IWQt:jr67i)t7eT-UґhI1bŨ+R ~=q݉2*ճڎgM>~%uCLZ ҕH-+ڷ*~xK2%HW)ŤT:xj㻨ä%D -(z)4h?vfQN1E)5;r p_ ]dv(EU χ>P7؛R>f?g|^vfY]mp{7۠Nw 1AWhڻC@QYߞ> ?ӷ_ut @^۷+os9w77E>_4*+Ag 9q״!F[ev|ķ'|s4??_3 S_rz?}pmԮ>`\{o*PhpmzdJ%%(mP%egt x[|q3\RC켾7\/JHb)vZ*Ѕ1*7eNKi6\_;SMN4ґzg,4(a|m͆>TC@(%XwY BVl!zMω6{Gbzo%SZ`$F3Z hGEIU@拘\@RMVa8EmzLr=:ܕ)aN" =Xz B8h11Zd4jYЂ~G"6Y^߯E6djmRJQivp#WRjh ѡtTPafSc,c1mFM#d<]U~b1*J!J8 }!iW\ϛbΠz T,Jv>bSw|B!wysT/zr}<:RI0j$7ܯXc;M)ɇ5ݜ|ODLcO1Ǒ֑5~F_ko2dT&ϕcHnNʗ VCJ ֪/-"$U[8c\ki_ZhI5; dT jGy(y3@f2i~ :jobE,N)(J892r$XT5`gY;!JPA/uBڂ4hvՕ/H!8I]ΰqUuAT"Fzm>ՎGA3yǒkBBmCH T'%Cࠄ2zNYh"1ZۃBIrZQ8b!z_4yZȠΌ>H+^ܡE{łTM%^16!9h>3(ԃ,071_r0=bi'oOeu V'dӍ_ڜ6/]D]ebG<.} ]{ #Fˈ|uI! &@&@^Z@mDBMuj WH% =nD4ePExhMf W nk 9Ԭkdž@x@'^oBZڝkYfՌ.#8Xx3F5Q0 td+!x݄i63+I@̴lIU ה!P?Amj DQ񨈬wM`Qy8 PY|,8+V blWNԌj-Ѫ %^!gѝ5MVU5@ jfMá ǧޚདྷPE3&BZ&7k! 2$A}>DC%l0'\qjh}0'<pwg#X"wi6+v̀ ]@H776@gejOak0gBggUCŨծ[stQ3LHY#h4v˨ <ӣOʌ4N"väDy ؓƵJC6G=T3ʍڛ(`fr5wn*2TA[TPZ 5diK"xHXoެŰc6Ձ'~~+Fd[RdQ'w d!g1Cʟu7(oV1Z8\ \U -:Q\jGn:T,~ xi,mTѵYcR=5 ޘ[Vj|ҡƚ5AN%_#Lzl5dQ Tk5>Q?MɁ{Bw *6tLP3xs9mF] z!@6(=< VAP4kMyQi(-\i䋑Q+Bjw@S &k|@?ѬFDp*qKކRZGvЭwA< \T*M .Fgr\U-1˱XTIKAH 8iBeN V3f !KkBE9:W-]1wE%2L0H5nzF\^ܲl[w{\a3k |)u 52j"v \Bq?Z ߵEhoٟ~7=Gr vM7W Z9W&|?X~=~:9?n78]zO^^{G:(ih~ǟGV5}߯ƧFՏNOnLxv{KO}>^lW=_=2N5Zϗ*P-E]5rpbhm:v'tQ@o dlxg-N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@' '͒@sO 2 )-:'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q:<*f$'G8- !O=ޤ(fr'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q]'j+ਖܴ'P@|Q @Q8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@o z[.x)z\{_&- Goc\b!,Ÿh6_o\ڡ[ /><-bbZb}2ޟ(կNWL u+ճRKWB~zJcqѕ]y=%DW j)thw!CWln1tpY ]-xt(;BW ]..+kRڣ+FE]E֮p+ j1th);]1J텮 ]9"Yjg-mrNWlk HW^)}o:߷4fP^[1{ӽmq-q]iRԙSgoWWІ<ܾǭ WOM8Ut»Hw~yxwvf{vkڅ:_q7.7Go. dGpašnwmH_aSr~*wdRu_6r)D \K#(y>H8 tht܎/7a> {\\WG@`A:pm9P`5[Ƣc\Z\50)ú,]w]@ߟ&ĪO=ÅK)}+(>b\?=X$V}Hk L^l.?]ՠwz8_MXҚ_(J;_c%è^h[PFW$X1Z"~9X#r<X0taE$6#Vu*ԯrU~<.Y{ɱV5Q -9+RSbY~OreӅa#*(`FB Dc]PNM" DIr-$rճ IjqR>kt-Hn>#\pȥU}_rV_W_\ZT?#J?sf \*QKUQ|_#\iJ`BD.U]%*k+R-~4LNRa?*Ė"U*y;3/_^??PSrv" ݈XX$UMQ#5,N_kNzV0|`͇:IR0/?ujl]y Vب锠;k$L;ƅqXdɫt1f^ya;.%$_)xhVxwdw3϶ce]|m{k ?,f!w~4s=wk[a`s o&waIu%GH|R5۸(%+q^Pjr5ږc~ofvW:fv+0u.r˶ڿ?C3[ްݎ]o[OR, -fvx3޷\AiΝ p=퐺/6ZK]F Ҫ&鲏3(OdUSϘ8P=d:FY)0‚S/Snm"~6(71|W-ͺ(w[VrߌoGUXSv+ІAUU.E}̧M5׃a+v<Vػ}|=|/RhR6e'֚%T*mYY1(#{hfTm,Z6MF7k9&}te'6b;ig[Pl7eSu`BZROIo!3=/6EFv5ViR?&vg5wG ' EH䑏PpX]썯U 3* mX,~$?2vQaE*PW0HpƂT0WvogJRVWÕLQ~}ۗ+EOEa:izgbˉ"1JtPkX* e0W:hA䩠c"6u#÷*OL=uwƷ,k-ߚqOC4e{lljX wQ7 bKub-dZ_g"vZI_XX1^naxMw|, 8 NENn0*e_Gb/P R]l!vY5GQŌ.ӡW14 3(Gge."|a|m^y3sxW)ga$h|awyuc]6I`my~Q\RL<,K "=`^/2HF `hQ*^.RsTΈُtlX3Bb!E= of|a,ps棲}&Gú FlE,2`xZʥa:D=I4H8elЩָVE]:bSE#iHƞQk % 6NTPe) P̴x*3bg#g?bدgsAmܣvx"fYʨJ.Ɠ2p3x0i"`vY˼lBg ,d 1Kyi5> 'HQ#ȁQ1092'wa6\143"{DihWd[:P*5[`'Cmp)LO) 9&e#}VD #$$8: Ą3[XA`IJ'0Ï̥*R[Ųh|ٸP\dq.Z+l :kXx#5^pCcJcBЈ"=.qǡx3!; Oa;<+:W{^o8^5h\ z]*8#_^~D^qE`h: G?>Q#K@G?VQ6xE}GKNB *"0E+ōRȗUR*||Y2 9 C!fHM> Z{1[,0V>7֑ Ȉ7$2Sm٩voD'Z=VY꒦W TDWIpk-Iiիď<ϱul6>V[Ox\56́ErKG B`U.JLCɓWB좜ŮB0),"/n3KH.0!4e_(*6o!IiQV, 4dɬUVA )WOו^:ki+17LzrmUzlApAEE4ӑLF[91*&wbI.נ.*sVeq:&kZ+6`۶7S_C ozwcEN HB QLM2vwzV!ʭWкSn{qe7NFC{6e-=wkzjĩCk-wl20՘|,Atz(#/!}m;"iȥBb͗c}Q*_Q |"o`:ٽ*YI\F}7H}[N<)G2"ɢ!O̳ӣsM}{l!?g"|fsS֞2mȨ(Qb4`trBj$#JGS~QVO!aR/1",DDꥦ$EV 1̞ltάqt6ڔ)wR7iP?Nv4 ؉gCWҎ_m[>/W\v@/N 0?SZ('!D3 ?̲Υ s0a ' pDHܰq =0<0 Ƙxˑ ))6P壔V(h#x ,:0ST8 y Қ)RRA8)l(c;Ja8#|ݕM]uO.Sw-v@vyAZ3+c<$5t^(% 2(7K2hSEY<3ͯ|=#KC@%(HoB8A;l,T*!(hD=dA7J|y/! # "K1ot9Y #$pr}XcOi,eDKq/٤,󱜍WºZlW#␔}D=J)m ~!>̤݉/$M]&ø>P5JP-1++FփɈZ8eSaltPD:zwD\'Xwa\;˩qIBѴ&^frgFU*%568;T w_:SUUHE:o특Tk|b<*%èbz8_eb3R߻Kq൳$A;ؗ5Z)R&%O I)Z 52UAyɕ΂+*- {(ۙŭRq8Ȣ_.kG5Q;Ȕmە}Kンb9h֊mPm$L`J9&=E :M *8ք&x&9pvn9mߘzQ]>@ۙ0{2a8Ӫr˺& VHXI#AIZZ$h1i )mwh@Z%עew]+Z.SacG5*iC)2ܪĈF #Jo' xfNyAYC.1DR |#dH ;#a5u_=&(]?n6ɑ)页4 ާ+l5Y>=:٣UEk4s`I^#╰J[, ZX磋YӬ$)6q9*Wr)IJm&YO9t0h` Tۉk^jkUwQdL &dZi,<&rkܨR(o8:DT,T@\(!}ړi ޯB[x3s($ԕ]*wचY^̳ǐ}tFU$XAQ˗<;`^K//lO×rQ܌\;\~z .fT2qŽW(ep*AG^Y!.3>O-;:=G?Lkx`ぼI]ns$tr3ou _ȳeas3%oa}TIJ!0e\ϵьUA&^) 57VBTpCkMR҂b ▢U2盈Dq0ǜsGu*y q#w@"OLX($($]8d.Ĥ\vo%_s3LC wIk׃=U1Thb85U r=;q讫azr$pnuMfzyZQ}rCԠ1}_wrlx_W/cɹFhrVhnPhk4!7|*z,sQ\Uuz;[6Pb?UKqi+E2Y%\g&ϫ4yG0ɸUU8˦D9(~ƽ8:RXꢾǿMF۳LkA2'N#qVU)n'֬RLIU$`{x3Y"2 i4%ԝKn'Ҹj__J*5õQT5EcKFFiWwYcGk3F<Ϭʶ1kxeR0O pECVIZsHzpHyܪ"&OFjD)0NjWJ*!Mz!,v"[,Eϼ;D(|HY |Ec l.2KuAD,y$ ZZA^ KruIW.@m8$tǐh̚DD4Bzr̀.DoS ʑpEVF!2Υl $I*RȈ1#sgZiSu#aHt[ ߚI+t'}Zx (MnO(h&A=AP}BϾ#ڟ&*wm)АH9w le&N!"kQ^`]Ondܼ.|SߚM|\/ ?>ǻrߜa܊.aB<+.opGLݸȕ{oGV1x?~ue{l9}[QaQ>gqEĚh|Mw7ii1"BQJ&-n|MO?{ v&9 ,PmJKQN+v&[A{n̠=Ս&DR (.\R)1nUDhp `D,H5<ƛD'sQㄈӮ&po\BEJ{}u^_Ou pCydȄ@Ak<@QMf6Qk, Bio@eBĉ@&80$RL6ĤWܢP%Ol"a ^9 WiUד.#9Ah,)*,dYRy>0T/$vӯxxe QR:\k K*u$vcg=vcKGcTK\ Hc!ybDe0D5%29@ ].h~̀qK~~>ww?]A=AL^K9Yz85>c_L)̖_dg~?J'1Rɀ<AC`KŌM0-P EGK4+1@CΪ8Zs&)iMHvT.:xs}SoqPr}݋/vW;׼D: &jKLH`" hÌqk`$U;mDx';1&>;Y<YPBn|7{oҼ&,Jf/7ݣŸwY -҃:Δَ]vyz̋~tXDrAEEm$Es31Qk:41MȄo(B 5ii҃&J1.$ 'wU Նbc dqX5c>(ݹpz-txuOxU9 TLL.94S;ΡJΡOs(Ti~@ ֌ι/Z2*S)MRѺρL2ӗe]]e*TWRSceSu 6c frBj)mWWN]=Iu$ 9o-]V.'OsO7k 5}aLDJ~OǛ?U}i˹D${FO; wY+A: %֔ r(J:SxەtRwHH]!N!SCRN]=AueFej]]e*TW2j}(*S 0TTv'aIU_DLcv4l'+ ns)RG?2(ʤ¨T:M)m&m]㓏6#x܄<<^?[Yj~NS!jӣyM-;3/3,1 xIrYtObNZqh2rOx > FL\Jw#b4cDWu:Ft#]LjƏ<i1:Ft*w5?]LjcD1u:Ft#]Ljβ\۲TuzǮcD1u:Ft9)8ͧLRlQS⽺zDj%e t\WtmVeJ>!Yu$۪z벿X uܓWBږWaꄗV$&V[m6|sl, >`B.FD cSF:&ܫ[ٹ}sfdY5B2@ۙ0{2aP{4Lz&k,i%RT2YcY M|ECf 2g Q!eJA1$x AOcEzX7qI]̮ǹ_\gNr;Dr ^%+/PAH7!M1׆={ [ĖZVnW :gC`noDߝ(ݼA<Ġz?L 8K[ H2OZ~j _?l?Gژ V=cO ~0/ r12hԸϬ5'5ԠIgcR'ԳIO`64>]2(6c5rLEt@e>.nVr2]>T-q[UJfv>e1~i#To>Z Y쟗-"?M\^ӤoRo-Qks[7ؐZ7$K͖P!aTRfk=^Y/^wvTҞ i26DE*kT!j_*זLT,,w+{Fu5d}ʆOlqs]/ 3-_Ω0[[b)[uq읹WrF /.oFc{%*ނ5*d#\Ğ=iBHDk{JBZ% T]L> D%QJڇS}Q"ƛkf R5>J+N^=uvϝO?>H5]&4OmIPzݸ9MuiQǢN hQ$\LF&L)%.b<((Qkh(c^+YP@,O,Y+#!wM:g; bLNn%Ϗ|8e12dc '5R!6K''T&wJ#E$leH&m%eQkxV ,vL;gs\>"h֛̄q~e޼LǺ.ԭ;'e}xezKr]srIbzJ|Uyf&#\F4G< 6C؜6k ClPb6+bl )R6*y =}N`>=*>m m斾N';+QNlD%(:5=ݺ-rwvIۿv%vQiQ>`4@|9ϫ^ [N͓bU~ݣ??'_kU>i=ß>BׯIw09yOjsxsx'D.ި 9ʖbVc;={%P4PFM*0k>gt@ȁe3ƈHM݈!@Jɓ=XiK e/3ڣBYׄE^2t(o,$-C*t=jI#Be(eHĎ1s 7 _]h6[]y/amWfY*~~-q` wIV~]>GĚnU4Is-ʐk9RV~zzD8"T]^q =00H, T%Y">+~ vkIԡ55:.m8#&)6|>n(_nFߊ|, %6^fW <[c2X_rr*Pa,`6] RO73^zBO<b1R.V$=i0Y>tOR< OGUOXymGA+救Xd4*yڐ %GjST({oѯ1ֻb?  mIkn '.@⤇ͬya<:@B9 w6v]b=?IM9ݭ sjugL'2w8.5?˓c]@~2sD3bo\')ٻ߸,i1Fb/1:^fKg"هfl ,ŏu|ޫw;ԇoGSU2ڝ(bf)F'Ah$>@M}ԬMX $>UѸt55/ӛҡ/74Đozl$ Xhjc&W2 qRF)kTKyg@8^gWx/?M/F0+.F 0Vv? M5?$C9_7[-U Un>ָ|cFPoJp=]o.}{8~l>lm_ӛcL" 맇a[އw/l 뿷9v2OK*kD vIKf6`]xfHu"2vO[LE5MToTE ~as]ʚ@a,6ZlV /o<ñWq#K4 : <0˨WV)ͽBrg"ŘpdҊ蘧H$M ⧼ۛD_57y]\PU(J)d=i~iO!]󛱙\2l!sرq Q:h=V9iN*TӚt;jմ&iMQ*f ;:,C*J`b,R1R z.x燘s v59^`<,frOh)'T+ąUSaQ#w*Siα(P82JT, vVRHdjX/}>ZOݮ s鮒K]kߋ3?V"lR`%a-t7jr"^-̽ny*9OFEz:z`4esr+ߟih=/E¥Pr9@{68uV,UW*c.+Zo_IFKf6bWnyG`kH@ċ(?Q@;< p<[7-ŏ̧ t].Ǔ c̱|>6|simI 9Zr;Af! {{hWmO/\:Ka|1u#vZshC vY4ajK^ ẃf5O1VF!T:Pi_M?UvKB!O0p_PJo%de ^$23X&x]d)&ז!D!e!x0k5f,`ZFL&ZOHK6j6T$iaĜ5q8 ⤡"FT a){KyZ~F[3nB^RME?dܖ2RfN,*JעbvQbJEaS}9=|q{i3SJ`0]Ic9΃ E"lLv *! I(6G6=3vZY-ބ$[;LDs(@I*M-/E (݅Uݟf^yz k5**ߴ(t_'#a(J#Iᵄr*c$: ԎkINQv~JYc[_w?VP9>ԍxq?g_K}NwJLe3{ n?wi+7NM(kϦ+}Zj:oLRkv'TI9 bضgn[$Kw-/CYco>JDoכֽZ^FJEkwntK{~9Q֫Gw4\mtkGg [Ss[=foιo>mVF06y퇚5;$ܺf=MjX`{*ܗ VAJU^XUD>l`ho*%/t2uJ(S)#UzCW -|yB)HW\x' ĕ*QZ&NW %W]@55ۺM[멜ɧݚU_8*82/9!/^g鼺CoapV7RcJCdߍ++|4.?=+K3&He,?300_$4-ĕI_6<_N8,>|񫼚▱'%>a aG]?<.01fZL,6uV\ʂ{嚗g}v̊J߾XZ0ʽ,!5Q j~4X>4> 5q'MLybC\2)np09sH[ ΉuXI\F|K.GZ(ڜ-rm|SaS:u&Oq4ȴEL nL N7-$kz`Zx9ƸGt3'ZyR tJi{up$W<J<3DZQ%zDW8'7ZyJ(׶؁NWzϡo0XV: -{&: %ݢ+}]遮z"}XU+Q_*UtPꁮ^"]Q$=+L ]%ZNW %%`.zCW nV] Ic$dAu ]ZI P9K+_sRH"K"R7tݘç|D,m4U7R%ոJpB@/ DŽ1*'q'e(mԼjX~' G(wp)BSϘ8P=tHbaJ:X'jq@Exlܼ'{|Y'߫?Z 0T⪻m ; x[`}`D=*!/e|%ZqAEW ]Fq\d9C:vT*5v/BݪDW*t ](ADIإ H{, θ<z(Pk ]k4PU+(*ƞeȅ4,P Ap:J˅ Q"EDaf (rQlEZ]! b+ jxT~8!,,I(J1,,VVBm >%y>4:a}6jfkmSўK%R>skΨv_%ăHTEO0Z{9'\A-r,ʈ4uVQ(֢dJH1 N9ťoW12#XP1g&58=Y5.lmfɅ˅ \xV.Wq~$͐{V+n>.;!7xT̿qV"\CÂS#I"C: ꈝ(iA7$exDTH' $&e) fZD"rRحv:k*fg$ضfGWX=Hf<T3﬌)1QxF8z Q$tWN;+߲] 6r!c2Hh5> 'HQ0G:HcZRޅlC"Ʈ0bkXF-3"q`ĭ6z`m A^1t2$qGX"pMi@1)fV1DB2Ѝg($^2XobA% *$53Zff/ ^,ͻdlmˋe^/mKS6n k ҔZ{#5^pCcJcBЈ"/‡͎ċ8><-6U|-#wZ{q ~fWqG˙r( p)噽E17G< rD-*&U n/qKͳ9]qs<Kn)o R1"#%v(` 5u"A*rʹwsDg=%*Eh;_K(b$_SWl@꼾~TfZkUZۤ%ގkrZ>HgN^80nzZEqJnnzEY{6ݞZ1;OԚMV޵6r$Ovp~w-`%9|9è~i2)izHɡ(j(QҬ驪_u=>y R+ nxtutyz^I/lvOl' \y뷟vޙj/kll1󵑛a6nouus3o[yLlQ(B)מ c㯆nj[?-inyJ'Y~/'WU9Z?i{&OE-U6on*wL r\MN~ݖ懛^ ^|S*3+gle֠?c6IiBI@bEq‹訐\P\QN(UMA%kQKMѹ(CTTrEHIe& lOT{3FΖԳ@9MnJ@cDJ>,ux̏gqO'GlN<;@O}Зk<~ˆl`.bεₑs4c U;GHro *//xP bx r"LE}fsTʧ\^FUxPJ٪lW O)S.H"I@^䜒SIks`=ŠL,Eٳb荜$0hsʫbΖ|'xz\(WuߟMQuW.o&3hV?WxdΈac\=t:di3&2yM$h%)O?g:;өz4/@ *-dBH"h dTr01OuϞ-adNB|''{%^x9|aw8UHHjr%CUVA0GK䅳&$bA}^A?\B{)> !j4IE@ySBy~~J1/n-ǎ2mUK{mm=a7!(Ea+@[ލp+mmNW)*/]2bRs>99?A 3gIXB롹sN˞t6XL2f*ұ@[lu9Ig, 6Yf4UH=[x0^+}x4ͮʭvrU:^D/àtYt6YޣI2rҶVsJY("鸬#7xߕX#vr9%.uJI,Kwb7%JW90fOa_{\k:hܜ} (!C2=x+*CzVlN4ɍӬ:6/=dMcX"6:JאEH-C5i-;kZC˚i-{hʬH#ʤ(5c(1hUx+$ł-V+7pUtt.-vƩqˀ] " w<XQׅUHE95-%*{ۢ qxN3pNt;z vh bF1D,HGI+QI\^S$0L^(D! y+:2>'(g!%$/Vcԁ<6ei뉒Rߍz#gKyM(bwyY)9ak *x,y$TU,e! 3%&cDo?cMo c~ClĪP[ump q9Z[tB^8_et4mf˙z9K% )2+5njG62:Gd$71f\}KyږX] eYA]Q#b!"<fĐ"d%bJQR_Zr:K >tQMqlrKE#Jo*GiUic ʭ5~4m^Vwo]~n{i220/*Ro#|a5(3ܹL_=_i u([#j6ZR[mIL&p۽Tɬ<:9=6om4'm#W&MWA`@=€R׎Cw:F:|~!S@>8 m`hr`K9k֒lQ' Vf;[^#-sRNP 8Ģ-^6QBA>U\덜C+(H|z7fz5V`k Lť,K]_Ol-鬭*(zY_xݻEVݻZ#/4RbQ͟"_K?L?XEp}_CvyxW2 )k벅Yrk7ȹyaaCZo P7Կd޺jKuz@|?|,Z9/ڕӁ>~X|~ޏX^q|~|2Y`VWGel2_(|џ~>gO4*FSUaF)1tCJ Tjvv`{﵌t-g/=~}3C-1[Z)~d)Y ^o}a⤉8`斫/LS hZo:{txؙNZN+h11kbLA 6hJlXࡀH{K^b)r+\RgjBϋN=胏$\ɔAZ MPyVѦ|ꃏkAKTdx[cF:v0axCN䫲GT4ilD4\ *FX;0Vw/D˛ì!T8/(2Ul@QJQIDoLDpՂHk8AI@I9{FD^^[#7}[Eת;t|`[/E#,K-Iy֗CVu$ ArB!@_Yy5 $)拗'Eu"Yi:ONH8Ka<%!hs7ơe. uXo dQBQ ^ x/+D sbptoârg:Ԛo}7^Eϕ&+gRx*Y1DIle|>S؁JCU5Vnp grRH'e^'_Pl6Ҡ0|o.kK&lb7^l;6I-tm4^D*bc <ۖ*ߜ$bR؎(FKm:}Ocm ;O:^ˍ7Ts:IҮd[JnD%BAJ%2Vr>aѰj۾(t6K*Q;F@:zID£,2 dL=qDžc$/ʛ 6/BǔdQD弓ߨMDCÃs)9k [dM=E~CK C]f~i&g6kͬ H1(o|L$ rJj$D~E9HIA1뛜Icnsv=v8.= {uG8-\ ɟGV=l4'%2k7e ?AUa>ϻՕRXRKwsҭ#Oyt[JkR^x),Lٻ6r,W Mɼl֝tX`&"K'N>V%۲d*n ?G"[u&aZf8 {/ZɽJ`XVtb 8PIVUk;-B{gɊBoc,>jg&녯 p]H&5r$lN(i{S4UE rkhe c9 !ap:R0)[*|Pmzlp~5͐|ΟN6pMi[14#9;40!1Cywy'qh&)f= Gr;h P%Nhk=֋`eX c̢(p:ߗ(!j2 C۩_!cFYN;[B7ƽu(mda/hܱ_38X+yB~-{d8s#kᬾ>q!ůev8ɗ/B6< Ša,`=MEL^ǽքhk^kB"q\L^'DUݽN(lŽ&aMO ny6pC핮6CK ]mWz8]6t9]%BԆ\BW U+@QCWHW 5+, ]% JtRZuJ(i]n*fQW .uVUBhCWHW,mkDWXkRb#^gbZyJ(5nΛ5UU]*J(U3wut |_Xj741|;;Y(6f5N 6m_~p"Gj 3yHgmZ!1_|sG*%F"+$Ft1UBK*ΜP&ǡ+-4b 7]mW6Bc: %] 7tئ 8޻3\ogp3WJ:@"RIjDW!QJpuu+JhiUBICWCWsJX*6tJQJh{ PM:bBKkDW 0U︫+j3qĒVRֆF.t+a37tut%P ~mcEbߟnPjhZXڐt+kh?Ǫ(I';Dk^ )BKpqm*tP6 JI$]`룮gU[UB+J:@V׈Y`Hp ]%3v %oՏCWMOZHQ6-.7-.7C+DWd" ]=Udm jBW -<]%6tutE׈06t ]Jhz,؞JݨC++FeUBU* ]"]1DA9Jp9 ]%TJM:@Z!^'uW7KjZVyJ(2ճЕ@,L[V.tmLImPnL•>uihZ6*D6Im`i !u}fp1fhB7CYtEzt40kn,5;:K3&#2:yt|epiU/L\828> y=$g $&?lv -S|;в(0_e~~'=qJyk>uǷ :Rl0eMt/dFDⱦL0JP3b…g!z!XFE%HdgI[ Gj*,U! =ʊ_ ](+)#*(!,F(Y/Ǘ>z5m8:;菏?s 0vqeCsH?˔)nxFC.Jl>]_Be._xۃM=u 9{T5NTxen)dn,MԿ՗~Ƶ43fa3:ƽb0C;Z4QR+n:R{f[A-ÛgMeƸXX4bپux q]WrhtῇmP6*8ʝ7\ 3f4j|>#,RFXp4Yy 0oHmy7)w%?u,}x7NWrolaa]_'@!8}=lϥ؞_|%oeP{=0{l{{Oc?K]kq&.ڰ [t1k^TX'pcU;l-"]x lH!R`+uN*?{ݲxZx7w~[~?glwW@K¶y}+]3ϗoӤW }lťႩ|<}\?Cq͛# U]fyRsR&#H|41-㶖&ǣ SBp10Hv/^Ρ[QlPqF P¢ 7kӖ1꧌E*'m k [6eGRo S_J Y0M@EaE3cnaȝKQ[%%oc&Zf=~#QתA};G#n]!@Jw 4B\:yXE"xc< 48xS)#tW:xCZXz[ O AHRVa CN'VcIDk1Ml(Fs+ f{sCP|B-(R~go\Ζ](x;sb&s)z&v+ALtX5iaZ`-gYO-9^w3I<#;A1Vc'(-nD^ . "ʁ[F:iG IM Q;7 qfHB9Uaawԝ4@Anc\޲Vb'K7Zz߫Jn9+4YG%Zo4z0uwQaINݓRr6éiNl|+/x0YHE#Gse+("ǢHS))3 %ZfR>RLzSNq飶ěU`̙p̸383θpgH.x\HQÅOʅ3DlqRǫ!7iww';Qwrv;؊XdKt{pʦ-e"C: ][sr+<$5n\rRT*R@u,S:]OcHJ,^D %d["A7䜂8j5j14ҪM<;BU" r؃bi^'`Zbq#ݶdNH35[b`$Օ |E:9E# Y!aѬ[ SE8(6J~S6 .>Ɓ`춇=-y.0El}jQq[fbmU0Τud9Ex\N.` r"BpbPGQٙ)8P; x=nn[mTU'n\ljvQvqVRQ6qJ.h $N A:fg&c( 0RhR1Vǡla&._KGcOKkyz|-zq#~U̽1rFA7F?QWOtUcěm[d'415<ܙR*H޻- D<;jp ˺ wBd=J}C` @(°.P\|_v#s3csY4G@s@Q?f)hkLs>egPlFWZ2)@(soSeٰ0lB/yLb3/,6h}%f=\ɕzYF˽lJGH6 !cnXû)ۋKn}_^l2vb:#r (+v %XB THƑV TLe TY]@+@-4PJI&撸xc¶h]2W֦+PYOZ Tq l!͡FUOh2ePȽe7xqOϺ,WٙDY0(ߖaMOh˫stn;ۛl>ն٧rZYۜ+>8I۬U7釮p1PKնߦںI") N~.dXMU'UEuឹQ5=S3/{k>- ,7)^/O.5Xv=^71Ѕϡ$wWRBu8`P£Z}]؟I ܳs : ~pog7ɛw=4Ҧzr:#8'ظ$d9pBnm~)tWyaOfLnsLON@NFFwۧ9nliNF)vf'_ym02!,Ϛ߫\Nce7kw%Znin5tߨ11ğ?)]F~`0mO/$j+é%w{*mc)}]=ņjHW#!X#B͞jڳnU9~FeTҭ&6)>h*ĒI b3&Ξ]ߔNc+0xu_}>?s&;Im'b*ȄC=w @=RZ3 e)l90x39,~uFpra4 o`I1, / y_NES'Gu$s.NGvPÀ§㜍/2f!RC^++ʩ¹*RfJfj4Y lgaF sٕ^\<|Ma2b?}D]nm*: ec};&)NNQErD֭5u 0rv)g-fyA[|A9.DR kIV @W|6d F3ip=E[% 1Ak-wiAzQjDf Eu;AmJ,נGLO(} %Գ@<p⳿!xSoqg߹J61KNsŬFsMT5 TUMA~W=!!!S)*}-жT1B&'b$Φ,iX`_ʴ?[?eϯ"e6r_.oE|ҮqJװ\SP@- }G>6`/7z4-!Fxv"N lzF]Ԉ޽lZ:)>R{}~5kAkL}yuWˏ7Ld^O=~)Bİkto1v?AKvcȷ=/-1dRn+Vㅊ/2Im&?*r[\ӧ#sa;8_wZWۥ=a,2eGd2&ET/2[Ջuܚ^;ú.sbo 9|ropx$%1(|H(/|[L2:S" 6xm!l\kbj2ѫ1sz~bQ8(jvs@#&$"+s '}v?k+n^oNPk DC6Xyą'oN9xGGDM6}Fǖw}3vWqZ:of;g wln/^׌aB +Et9 .QcFq2:dsr&zbG >H3t62 P"VNHEPŤw([kO*s񎑏wB߼RnV:wJv+KQstf*PQ*֢P7 |Jc[k7XHu]9vKT(pTuXy]߈zkMy:?_Hkyq:Mg NX2K E1AXe5;=H)BZ̎s¨BTs}'K WMK+F؃|ѰjJjU:ĒZ|!TNPlP0& D$Kyݺ lg7 ='i_r זyת6S1(*z[1B7 ݬ|Krӟ՟M>ynN]!uhIXC^ -vڤO]l} gУ- 6pjuL}\a焵s;( ϿO=(b@)i"ZberP\J2*t1y0:T5]N#!Ӗ/_AǧK$ڑ¼ Vv)̣نw/4`& I0GPc~/C85P#ad>m&S&e[U!AEr"hA䴎*Pe :]MY{PeroY+Oy}I9^}L[[|,ד eVF~]~RmtU>/?7|?}jX}\_B>9s'O0K=Ȳtql 9!۷:W>vq\nO\4`%,FTkVrKrvI𧓞8nZh3D $ZڳM"Ou\xD:]>)oM)HGBЂ)4`h$v3vE]*&]z ْzx_]jŁ YAy!GjLб=b:<oN!'21BƆT^6\&N!$kQHo?9n ?tsڇ[M=5^90h|zz}wٮ.PaE׀P|be!VEq8 wnX/{[3eb{ZRa'AaS :#&B{tm- W6_ %㪚3_ӝٟLLr")_"Z:%o˙yZ=9\:)|"`|NɉtS)Qfe&W&@pZ AS!<7ٝuݯc!wE#t{Cч^.{V0ݼa)u?v00X4I{!P ,FpD* Q TJ d 4&ghdžMrHPJSY *d'U6 )q `QZvlhzmoW[*29EE,-|fɈ ÿR~Z~|1 ֔C")uɣecԖUwhyeI-@Nը"2(gBR۹z|s;8c]JF=$ yGjQ)1h)Z,5ZX l 񠣑 e"X[.$ Jb,5kg5.k&_fq?wf]~x?.la&]v#n05ٞ7bZz2#z3؟TVY\J^giuzX-fC\ex \eiuR*7W>M9p%m23VrN8&+Kr(/8ߝ*Q}+sG%1\.AH"U~]`:žtv\7iR#B%2&sVsATOUxY-b%A8լ0T{/gPj< ? XHҁe`;Qw {ܶ.-2 Ȼ,.>'\4WWjɽؿ:xmG?x$h<CCQWVzuRQBd.\/.%8 Zc7DZAI5(ӛ(>F'ƃߕ<yg00(|U*f{^+׽*͎>͠/Ui:Ovx?)-)ՒD)MrY&_W/:P1$_>⾰iҲ*4)9-?Z6JMͦfSiTl*m66JMͦfSy^ǽo! mz6=AMOЦ'h ;j򠶗׬a'x|{zעq<<y֬H#C'f)pLTP9u*֚W/TLF %&253s:%Qy`/hA鐱r\զŌ2EY&V=yyU%tᨸY%Z^0OwQxG=L6Éj/};wQ` 5ꗭ&RW=3(CRPqX"{zjzտZQ&Fz6[d"*0VYNTSG3/QQ<5c3)섒%4K&zA;4A1\Ά di ُgSu o< b'#k'Qe C4CbIE!&g&k]F,7]q(. ٝw r i%^E( O\߷[)Mo/ͥiwi;UUi)]qcҴioŷ k׍D89s)w}!5A:j{XWU4nyۋ6|wkZ]ɃIW;yP*nNzQߴcҝ|8tzzr64T5 kMUuDc}S ㌾?^;Ⱦ D+JbJ/@6a2Aj'm'CnR`CHԹ1Q'4I5d$W9JoW80u P:e_oJ)Sq1 CctibTI:[k)C= d \l B-75j'jέPA;&q O,C©ebG8#t!JbrYB  ]v,Qi1 9P/Ns4V i(˔ALB͔12J6, CkrSPǬ`Zkr(: pTt:8Uqe])v5K/G-Y52!yK3QZ $~ڋ@#c(bIQfȂ f6ZT0g;tȹ4o.ç%, Hv!yf_< 6c e4)dV0 1"t-/il.鿱3wȲ<ƃNE\B LhsHN?\F YDJ[D& 0L*mzmϛTVj j2+ ?Vޤistł #/մȍkoW߾>L{܋B&[|v4x)a\w]a2p=M[1ԚS־'jmټATj:MbϚ,CzVWwz7og$o+IWt/rc}Q G^ܥ+p?*%EC.(]"gAʫxu=qIbNdEށ xf1:A=:j{p3,ϖXh.c,2=<kh.Ve}a ,Hg%?ꇫ&$-✌N^-gyXbnoĕg8},8'/T Yi]B,s !8tދC_#^~H0g)V! Z,(%T#qXu ˮ~lW׽Oi? ^mIODy@אK2ΗT4DG%iy(׎[/DN߽K9sum/Nk O]wW[&=ճSaS2pq7Y{}=n|DŽ;`Z(jN~/KnZe9.N[Dn%0P/davH"nCn[u;z@8zW k>dC@".g^ٖ~'_e̵Zo/1qkrTXsˡVcn싇7|=pv{ Fʎzfx% \kA!A qB_Ktvs]b.`ҌIQ`VW}?X~1_,vc5mWdUcz$OKa?n.J,Y<2t2PեxRAm uAl-X:yVHGe=Ok2BEBTN29s:JW`Xf(-w b#PJnIWyؚ7ܴB7!xu\XP>K*c}1乷ܔޣ6Gl ܚHxU]j}թҭL#WImFWi1kD\靥=xg;<ժlJ>pK2hL>R:dd/)ep@ɿL:CXKs@}F]Z0,za#27&zzr 2ِc) :(+ Bu!x`|KU'By#cBcJ~}Afz^ˍ¶12r&'˹ϊR{ 2!;l6۷ 1J E@2 Qf  $$AHZ"Fj>@UmJ x=ioǒe/ICpdxq6CS3/l/(q(؀lUuwUe Pad?Ɖ1=6ðo\4埣eWCuDe7KI"WGhΗ>#( c 9eL:l&rGǗ c_#`^(85iCR*f0!-wQ(v//(0z 5뛂Lf fm:K9#PScS Wӡ §1Ȋ+!ap1S aC0YxrƋjr|^9[ =Ccݺ*9HԺoǣ"t3P]@ Z Ր:`"k;؈>t{J"&zj3Nktkt;j렮+D)aʳ I9ϯK,yO1ze"KU ʏ̵W@?ū_^\/^D]r~ :? KU1ס(4-A 4UlмQO&|!}m ml-6//ekX${~ZD=Re&5*n%+| ۼSk 37Dկ7̰r_:!sgv٩ğ{*8hAZ5%21IF^)lE)Ppc'=ábyy{=E# 6:`\܁rjh46ZSJ9gChN,k:ua4oIkPM,ƞyf,UM~Rw(57/\Bks~9ubR9ccfP&qx66}kJ;z2q[lm$/qmpFRJ1 pl&.rc@0ZAu^=ig cYrow,:VƼ;BwV3_wbm51PNhG{Ҏ#JMIԋ4\J*$ T? |Z9!;lT^8*~y鑳k3|70-6tz[!L l8yb[K5ч3(`d68`>+:S= eԭ~(cp(}@` ;ڡGϖ:g9i{.o Pu4T}V86㧓;ǗnZn/n;;w{opˊM'UqX6Pz[)TfsoaNOT h!~) _d?9Ff0yeX2 V1J`AqcȌǣV<{X{64]o{_.gC @ 'A^75*$Lry* s, xո8@_&nhl lZ3_HP&Sa(,O9~l$PTi§SfV.J_˳#%<@rFϵ,wJÃNG"S='w`ʇvfH*/R[Ƣc\܆z3c}nkMZnXf9wˌDteLrJ͙"ǜ M0&ZYBee`c`bҹV:F j 9ank@Wk=MrmZx6 ĮaI]%tę?uvTM9+bW a)]%tw]%lwij=, zJ : vsJ]%t ۧXNJ6`W|åLjHt۽)XMd}3&GX٬&L77wȢrIEVe~\PM=j<$g $&GzoVEg:)}ӳ6GH<֔ &\ "wYL,D/ب$Rŭ{`& ay(ɑgn]FW ~bǻ.ЏO-9?s r.-H#Nxni7ߴo7vM|ni7ߴoZa7vM|ךvM'S߳=܁a句+p?lw]s]~XjQBD E/=N[*phSqsYNDI*zڎ[m_$q7N0zׅ&0IURSCi3 ƅNzUQ>d7Y'709p1V` $2IDR>0tlpLs:ܫUÀP& -rH]% ]&媖BwT@g|jM^IQ` (;3\,D0E`]}p:ocj#-jN$3'5HF r)qYH2"HfCF LI)nv8T,//vϷtx '0f1{uX X!֎Fcu 13z0dIJ&SK\gF[+j_ӀN̕#z$`{84pQj2 i(P5Pix`,(3xn2MŃ4U:ioUڑ[|i,+oDŽ3TzPc4qZ I[cR z^iGjd@T{1[,AK1Vٚй[OZxt4~0oi RλWcWY!Vu'±юiΐh7*Qni7ߴo7vBH(wlڈg{| hZ5 ò[ 9fo}X [܁Iyø6_i{O7ulڅrUiz⾿ s@}mڽ$|v}Kb$Ed8߆I. {3n˵ۻ u%sl5٭gq%nd6т~nlpzA!zu'Ԧ;݈Mk.W*֜]5- FړvMc쒼05ݸ>hkP &mx`\QrPJoԭT iW ESSucNucRWŀ-Ós΄E$(ڥ2 $P#<"ުWD,Hw?0bi*OcD2Y+냉&eDDL ` Xy$RDkՆΆk\|N O .{wtŀg&2ﱝjױu)w'(kiQ9΁5R"!Hb)Im) LHgpƘ"h֩?eFтXI5Cbus^g}x<齃_ϺDC[kT) ,dYhOInEowK #![dQ`9SVX3d, XDGepGdWF D,Fг)$3ǩ Z&LNǤpݒ:S)V??#_.O 2|4*TL/m= 'FG6#X {&s);Vۜ2*"NB4f&#J={71zpVnxQwK#e96; %=Nh3NF%AXo˨VR& 1A}0A*1j 3ꄷ՜Sͥ FB*1ad0Y¢|dk0hB*gRbr𿠹o ~+ DA#).8՞0Q{*bY2lEDjvc"{&CDt0)RRA)DpC@ߵXZYh6n'qmܭOpjni`%>w eʜgIC_~|/Sĭ iuR Ҡgj{_ӽL2`{,v7$`,iԶ_~VKl$˔-{:ز٬SSU&&j Z 0c\4~cBEEoo\'>yxگ6Bsﮡ5|7/He(4&'c A] .qX$VDrһ*XtI[zW@g3丩@ 9*^ lQ$x( Bx́$ؘd:Sӓ4'd"69f#5&錧QlJ3.(2@rWd"b1qvؕ#b$ˊ0VN˛G$u!|ݜFpW{6Qmy%/:%-@LW_*Ø:l\U¡[i  &8?OO֘JnjKS$ّGMta+Yapqi*\t)l/o}_{B~m׻x6II"(tz 3x )9a gC}߇eXgyVKM4q\0c4ј L48U W8o\Y" qeK^Z+΋ w!`I.2Ƃ&!$* HWC( Ix9F<7 ' +\Qp^$+chc<',kW\EZU8h (%e@yrdK;>Cޟ۴=o̽T,M10ΫvR %MG6*1`$VVKQc`p}e7f>`)p#! z\qt5h GcB]s _omTDYD+=ո;o|lf̳7btR,k7.=z!HΖ `XhnS%0*U-J:@zmtaNmGvOYʧt#P@ dx"Zq( <1$IK.OVO&5 XN)lR@#y=T"r:C&*O- uVq N=wB>~IC<σP[pyLuCԳX2nv#)y'yqG;p~\F>̐xeo+ Hb'_w59$!g3~RA;,m[ Pks1cM=C?73k3N8sC5~Pats@BBFh#fO\T#? Ayqk3NC(wusgi779wMLB;mBP]>`.64_g%JTƄΝDgEVH:AJnw^[N壏6s;!^gimD"q&hР]rp!DrKtrxN3,)<*% i!ĩ\Q[.JY3DbKg,&Ύٙw&41%3 c,m^|^.mF "\gPWH,6}- Όg% PkI4-S6#ئYɡCoǕ,K &.z7r9_[yy)z? cns}oG ׸ NO·|4?͗ş}r˸ ͪ\|[|y\x3|Wk9]~mLа^O6+u3|ZwM|X<<[89@&||b|\ΖB$(* 'y wἚ4$G46 O>K .$A\}dGS:퓝w[>M3µ~F(xv`ZtU.r;]LlxZ}Ej#ta%SqaΰM=*]4i=r7; -x)(컻ӎbLî#)rg%i90 AxX AlO*W h~f{}d"XE s5p4{&@L`a-^#vHd)q,9410A}J bkie3MP0$ڰV #j<.p}zձSvgܵ) H* GyHDZ(<&ХukW";G]cբ7І{Q| o}]5eT/}ז{%9f!Lf1?hؒ-CfNNk~~z7/|z=J8C|p1$Nyxi'pǓ8lbGJY3=6- b̏mmlb.y\wSۖ/^69$ / f-m}n>zVvu|޻| ^koͷ,|oO_|]oC?*}xz6<=GZ!d_>9x?߃ b p@k@i3+_] CWW3@jVҕ !]pi1@S2H1+OW}e !ByYK>|z}enNmZ ?(Y뗫 vUћwZcyyO=^E.nV엤*8jp_ MygHؐ"_/]%D_ w OCMCN>=pۣc/{m* x>R;`GɪxdX>>U %ʛ;z~@Lç8 EO[=)oY-K#U~[2ĝK8M{eds$oUn:[wsҾd o1qj?O%܃ dn}=iOѺ'vT0PG'~9KɛtQ*k=ঠrLΚz>;(Ev\]kIs4PbJUYb¸rTkSUM=Sy0v:[;h.dt>3YY> ?'Zj jXqf[%k`DMԴA3j ]^u9E>d'9kct@.G>3pޝ+5CwJk kg3kkRSLj9вtYPd0vcF4CcvW|D\L5CёBE5䔔5# L{}1Չ&6fU݃6\esrʒ&S$s*0̿KoI 0kjA 1Ȏ ўF Cm.5뎼0Q) _f E US`ّ,ܢ >/Am>+h-S氾mmV Eeìa}VTbJ\A2hEgK'ǚژ[]KP\Zlhck +3\oe۸0`TL.,a}ƬG6ء`PѦ 8%Ԏ]SPPBhMA'SX869vx XٕU!\ "U+5IQ22l`_P&4$:3>I: d>b֨(WFoJ'm4_P^q3j,7p6L&jlU %Ә;>J͐jPo\?1u2ߺLzNB€("2] INbTytQlt,,x:M;c1PPJ S!< Wy(y3XGfk!\(VZ hrAN)(J892r$XT5@,Rl e 846ՕH!(.g-UuA"F_R} f1rոj0 !ѿ)`5R.!pPBY3'Y,k 4M4B@U+cŝ[ }FФi!:}p0('crn3RUNJYdFŘ@Qb i5 !pB.}oG Eg⅋Ώux^ ^n}-^6u6#X 3 o byPT8xiT M:V%H9@ҕkdAh2BCT< Dd!fGf1di*h MKе 2 XB;KRn ȋY T(ѣB- 1sPGmwu $$hLEQ{X1z,2m!%TmIJ X:u )ŕP3XU30bmG39O W/pwG]Nqmv~T&YLj`.;5B'eKa`{BNd4QG]CbƽQ^󤑲z hPAi7? AzV{2jk80)QC^";$u5\Qr#bosSvgr%5n 2TJ`QLYځ D!(!= `y(`c_XB|.w`E^1H"^Sm:X&@5?Y7;Yy9DyŰ*aL);ƌƈi)7rF1 11t \:'X:W![QٔJǨی ڂM+qҡƚ5AĦUJgۂNd &ci@r:{OkS=Fw {톣KsӸs ;[*>Y;Tڠ@4XAV -F3 _B\̸/6O34%0d*lp\,*w,r1\rAv*f-Ԥ3|I4M](IWKl$dipM((]U~rE>r앲7L0H5v[F7r/n>::rk1lvVkbB]dTrO}cwM5SG{#ʶgwqlfo8kh'P N"MڋH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 |@RB  z&@@8ȓ'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q=c'. - a9N ; 'st1)'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q=_'j[![nb@@k*'3ti $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@z>Nn7gָ`zpl5loW76 VOWM ZAd/F0ܸB=u@y^%ƥ߱q铡7zlY]0bj6NW@xW]YhAt5ǰܸ[@kS ]=G"=l +֞CW.Ў8] ]=CrW+j9jf)t5'OW%s+rΕAa`%qkA1z}N.ȶw:p?(Y뗫G+zs4kʾ2cnmi6 Up_Ú~̽q[?~< B_% ̿WG/. ߵ~9s+קnN ZjnׇMNдuMHE?gj ~;/5pj9=:h <i5L ~WwL@q]VSN=,#3/WN9wMoX%u7ͬɪѺ0Z3S  3+iU1+xW *m+B{gJY])ǰ;CWWUWWwϒ4sN˶FC5k!r={}.U9h@ #;sn} &uA &m@(۶k\?.C )Ct-c+{z;]Jя#]YgRV'ߓw?Bwj}Ԣ+]2NOW> ԽvbuS%o T]z`1!"u]+`\|3˕]!࣏7HWW+D)۶xOWBWY+CWخ NWlOWgHWzA뙫+tEhn;]J+{:CRU]!`á3tEp ]Z-NWΐ40mY+t3P1źxǁ)uIY!F3@pO}^h93miBO1'M 躲-?=W&"yF1ԋ@ꅷ~oV# O_W~f<ǜ_3_'Vm{:5 %2je#?0C?#mI ||dxpu7[S[eEL矏~FWZꙕN/ f7wF׭ v7|oN[WϦ6j/;?יf3^iG" iR^^ k.yz>k$gs*&F娓HRzYdhӥE$+sֈV+8Tw>?9p#li -`Fn8XEjq|7LupL5]|ɣE_uT9]^\<}?r}F#l /Vn{i42Z~!ְsz#~Λu.Mv=i}KYӯţK5X?es/zzKL~U&WU1^W3T&wǻtp~~a7@~:yI` L_=&HVmǟ6t|jO{w,ή/~p..OrXp3].K³B np…Cy?JeO 3nWW%)0?  ؔ2 ɠ<{_4KŇdEyvq>.)Z}nҲ/}ۏ0/`9\KTbb*+԰&0~jh@w$s cTч^mv?M!D5n9p>B@Iw apuoMɠ#*nyAoLpm9@!iz2=ʦ1% -Bj7 OY-9 &a/,%g _BNms0k=1grwIߚx7o< uW=H9Nlzt_Ǹ,ƖoB_aYvd8~ ^bSXjuE~8PCZAԚ']1Ai7a(%d1=6Rp=шc%33hbV ]bptP)6l71&(K NP`>>zfwi'Oo~Q|D~-&Ԕ`>!;Y=Y.9iolx-g?e$aP-h8XQ+MB'9 h!J ,b s"!eFmMIh[[ԓVT\dKA /ifl ؜Vi krfP ʅ+DT)͐7a* &>Ll8 *? wؖ B/].F%483<;⊏և+Sp㊎QV=8!p66Qˠ%*&LAߎYW8"N 6؍yx\(ئZG[Xzϵp2` mYKEbAe*{Ϝb]uQ;ƿM (m| VD!H&e53jQxTTblo2N>O&ˆCQ4̈gĞwznlDPPd9D=lNB\rBd&4d)5ʈcH􍳒\D2♐B2li&si7 #5̈ydꀼ8զZɡ(EbϋZqF2W!`AefAF"H }GV^X)0L cb[q(PƇGyM#_ 4͂10ς^cw̑E/~n  9 _G?(QvG?Gں6bxUG2%Fy?l eK[AXab*8]8Om7#oX(^D*r'*`nK۟sʶǟ;PKn8Ō*] .e `zɊeLnv! eQ9, iI2xmBe2W`׃KMO4.M[$J0< q/??\|Ks_A=Lu5&bcS1kBp %x㐃ſ`-DgYt߇eI:%z0SڨIئ6g_ke,̍%a-!Cv ^5$'y\&Ҫ{c9GLsF~F9A:aZ8J@A#(.rD0&NG[]UW=U?aNc `͆rb=Dgv s'Î[j-+% o#ă8Zvj)RFY&D`IBÏ!Tg^IR匒U{^%Gvw ҅R^ɐc.3!K$R=T6qg;9)O_ۦt]ߪtwWKH*8y]JtĒuY!ݖh*iu[g32ynGmfmL[fm|nB<9JQ1L圃)@*DNHlQBoGv UbiMt` SE`= <yLbe M:e2T檷{'WƮC^Ek`?^|Wkւv2Nh&-'EfWwMO* 2c䱫 r̒CQ4at'8߄bs9`_b&$?G(MM(i;۔C]֒# LӓNq<)V ?!I%d cDQs _(Ǔ=ԫ De猲ND.D: kgsgiׅ9|m 5 GZ8-X]0In^us+eA6wխ'fw֔'6Xcۘ}=+L]RWԻ9*( u”]5^MΏ_|q+r^K=L-~,kc|b_xѭE(R^5eI_;;e,VOӼ纃)u'Y(_#m'i\֡j֙jk9P j fIgxR\9^x]oeirxL,W\emdK#MSXwOl!0Oq[Od\ҦP󶘢/XI&f|d))€0QЉnѱ~sclNRwJm:cK ogt;ɍw:16~iD3䨸 sTHf޳mlW|.B{"MMmHzqe%W8E!)Y%[)KqؠM93=ze ޸#l#d\nиc cc1uƔ'H^p Ik <!Dţ2ʮR+%.,꓇cWÓoxo5Dtk+Ȍ3UJ":chwdD 6`UPBKS1vp49 Yz:8I@RdU$\ʂh=$GB/HTri#"@?ɍğ] ',hqs<L$SlP@iZbC R@r!Z,[Q}kTN+b]T-VGCVꋔ-iD@/37no4aO'\ i(lX)%dK$,q#?wl|iÌW_|\KJZ{1X[$ X@չ{j9h.YB(~^ *? 3=\r=/.zP*yqڼo&U$E\e74} 9Ʊ7?Ga]cћT1Co-Wfb_k<a9o N*VNi3f؊ [w E}We-msTa[] ݟ(8FW9bKPc._)PYKz0ȼ 6UrZzG+ 9?d./QbH?W'=݂`8:9|Lx[ϋeO? *vUY뙔\=yէ >F:V@wŏ K,rͼ⎹!2s7AkJy1٣d g+7Ksgz1frP_Ko*[v;ʄTrR!HNN^ZZm8ѐǫ9keƧ?!OJU9b{:u"_h+/([) 3FUt 4<ߓtFz[)z 7Xn7 RP8͂[@k:{ '16bh\I>\8}N9g@ GԻ ugzd/D2J4WTD<g ZB! D]cv=*O/Ƴ5pd{Or;a 0=F (:!DB7H .IYyV{ka6j/-{ϫ܃ݸ?+R(/ 9XcdZbg7O_sY[_F92{_~1 !9%}"` <%Aasku hK-u@&M]=.;u*%[ٲZC &Gs_s؅D8FZ4QMKH֙Ձ: h5i fRT&gX-Z8"̤ &D<)5pz\17WW6𮲁mj,wdY*T <`\U@`t1K $RBJ+mTRAYY{j=SMBq)w# (B<ňzM>F3Q%hoURRsm,n!t:ik*O934s_D].p'b?fqKc>y}1C$<1BIm;ܢ-p׾|b˪g]2J7w}eTgHzh- CX^1)hBo^V>MH11WL* ܐL0K4$hG.( mijC#IhhÊ#qQ'RP0/SJCL1H|$ qRsf^"(aF'4c$и$rc|:iO4 W8VZ)R IYIPPHqtDT,TIƪ{C;{LE;oMG-{/Wi[x3shtK@8e 8%gB&SQ!,T (JE̵7ۆ|?8*)PN4 A%/%w2*'٪iia+G1 +5dI> =w֛q2rӏ\5o)2ulF;[ S=:/5*ĉ/ ۅ":?߾<_1ew߾~RQ0aN͍XgjW7 μO^1mgVmdkE@+~r3iY#\KM{Yz0}iUYcoP)Tvgb!uzikHIlvCQF,uD'rd9N8KHR@#UU)Pig'=aʊz>a뿋(=OAHZ< |9Q0AdNt:Upbú[}ۇ<`돞,U]nBwڴ3ݔY Ji\%׭/<byHcJ!,]4P802jNYWdI&6 ;z2q-T w9jǓ&(yb,!P4Pe>IITpnWOpcPႱFB& ("&OF+ը3e=5p/=0+4ʽ-v\27Ƽ;BV3o(|8KXX B+cL;δOq^Z.eI3*қ臣w ^f`ͺ8>ų:n9yWo/.E/φ\l_C8Ň+#o?}0h5.ӒX\,jHd*f)\ +:83=ߕ2֭=o1P Yо(\N{([xE-iޛiY74_K/*FQ[7N0ݨ⼺L.Fq|pկ?u:BȕT {< }xP 6WË O?}|ZmߘOG6i \dR%bG¿ Yp8}$ GV~Y'8M@HO#Q+C 'r$#6:"4 eܷ[&ڧV0;?i|Ŕ ;L1WWZq?%#m,Yj* p{v$Nʏ6QSBDJιC u! !a4Ӏ6*qD=Iᦅ}Pzk/ _8+\/;-eOcqmW(ǻXYi!VE4.rٛ֫ޛ+N3/G/+*̟~QaW?j, a_sAݣb maiZ;Ҵ4j{¦.Z^>ZPb#. Qx8RQ4  6` )h H%h`SFn\*f qÀQEGKڴ j! 4dDf Zs&) "YAnj MG'np\Ge5OJ1{c'x%'ҹ(Z65 U[Ih"G'(BD|i{2b:fX-vYv͎b7,U ܣٸ ,n( ZtPz+.?{g㶑$"  Tznw$Y#)Ʊ~UFC{FRҘ!oCOwU?l6;<Ѿ+5mӢ{MZl2mau*J ,Dۢpe{Bv}1C0d>,-!p&O3ENg( {QGs_@%+j]ڏtu_'|WƝz5O;0wvbj7U/iuFu*9D>G)8GQWYYk]֮c]E۩;UKJҕu=v1 uFW.V^WLpu%SE+F+u*]1}Q+Py"[!d,bZ}S*1jҚHW t4b\Աh(M{ptevb4b\i3H G] QWk6>VG'ڥr[_r]mF5ic'msi%Q7dlͩ6ggLHy @(>}Bwcv7vǰ*'0Ň,j)%vKUbVܾXY-ǹd+ T2'/'ۯT+tV0}Nj3lVYyZSfFvJڇg;Uy1YZ.oJM ZzzW//LYh4tL$5OJidiO~9t56YȺ}#"Dp:p <:DiesEzsR8Q1OPA3)鷐V"O3:e0 ~O5H(>*9ShkXNdK",*ˢwVsWilub.~:[{1>$`4La\zP}aJ5>␈3F5YGWeZ#r,=D]1c}HF3qu4v0dit!甈HW ,͸> RKSRhAߜmQnbc`4v=-tp;JQԴC%HW D+ui}oR{3jB4OYuEFhtŸQ}uŔf] QWةxnEDhQwŔrѕlz9~C+\;ީhzl;J׳G 䨫cC^銁;h=)u5@]IՊrplq}WKOI]m_(b5}3v.j[*]'ĪL|:),Ex1Ûկ7(.'f뫪z_x)o* IUk*Vd6ZLCX)L_82TZeÿh޽ke1g-PLon뽪'?MꣲDu.vO'[pX] jÉ rVylU+(O_j>Q( ~^-~nq^KyY>eQަl\NkXkOvwߓf|䫫9_$+tejidm1]7EIi ˷_>^o9zc ll41Ƶ.Ӣ{o){c)/Tm{c_ͩ>Tk>۪j:,mTyR۞Uuk&gWOiCX|7VՊay,|N_gcaL":Ɂs|;Eɉvh%-*Jӌg>u6G{5nzF,nװA}9Yܠ&|*]n;Q+f-fMd96́kGJ1ҵ⪘!H:vbda6_([?I@A}} V5{YfY}9mko)w?^8lCpւkpvXFa#z+;"3\N~8WxqjzrB-yղP=I˪.GLmU {cros s X{ kr>2~lݎ٭m7h6A^^您uΖY A XShҭ:_*?Rz{C!ْjtogJw>>[?X/g9*~pO]>7ڒvd(e?oM\-C6 P'kU6qp]ii#W29 d4sb,sM0qkT0x41}ս`ΔV0c=hއ!Z)EuŔg+PY &+ {Sū8I[~OyUUOaGYSNRl_ (0M7ejkuULf`6CD?9w%J}8@Ǽ]- Q/n/5}=Sǝq7˫O^kg͌_pN@y.ߜ6-`L'IOf8D#{Jd4}8iI[$ >MbsEzsR8Y}M&8/ [B̓ESz1- phA8"x0.F.:ѢBfL fuA)p XtŴ]WLii1GWaT5,VVwvU+Z':ډ2ԨcC銀QhtEvgH;JkF] PWR*mDR qBgLbZ+cjRҁ銁FWVR`uŔR_㕃銀gqޞ-]+bue9HW 1]eDs3ȴ]WLiԨʂt%%J;V[xVNHl"z"ʸ6^%b2emwQ4Ms銀=3foh2}SQWCjQ=%`2LpEWLvvu* e4"\V~Sult^Ot?=Q]Om:jEkEGjG){on+=ЃQhUDhtŸǢ+w]1%QWԕ4zWl F+E w]1ػ5Fʈto֛(-QWԕ(D+V2 7Aս(8S8D+u*]1W}Qw5H]oV%qKv*}ctr9a{!@;kxvit9i; rm2䈶hc{Pog TP^?azEB9uZfhp -C DIY;- ph9$F+.]9V Sq$tBtEl4b\ʾ)<93xtŸĢ+޿Ŕ!QW4 #jTדᚎߑhG:zGeteZʌ:6VXHWKPQWkl,bZ4}QJaG] PW1عhtŸĢ+%_]WLuTD+WC4+ubJ+zW ]1wŴ]WLƛ!xg{6m;`htEND3δ{]1+ Fv~R&.48Nh% "H !Mg̎h5@5͔֍Q(12]1qiﻮr1H]yaP4GW+ fZ-+]WaTT1v;U+\];ZhY;JٳG 쨫cCHW̋6]>]1¾)-$tEOuŸhcRQWCz Q銀GWhtEZ^WDiG] BW? ߫נٯỷly1j6KoV- ɏwIosg5%WTo =(ά{B׋GեsjRO/0- 鬸+DX?U˷TU%bEɲ?YՆ}]mZhc^mseȮ6]683C]?~'Fkg'~(0+jo>kpwY;yK-ed[IxOxּN\U+Rlfi٣:".H$w_OGTˁv!!5#at\5JH0oAWk*hݑs=οz>DȻWDE]d佞c6Hc/S_,(% XTy0lU"- U2kJlAi3oW!IVf}Ѯ_tqm0O*jJyW+J{]`:N>%sԦ0J?֐Z:?2 ,}ꀪBPVH# i,WTt`t"!;;C F**petl)!K>wiok H`tgT栩BAZDmI"eD)/<`xкIȽ+A92 ^oRRiRբt6͜Tr,TҭRip:)J'e )@B{jZ=E\z !3e31KYOuLeRNt& e ˝;ᑌJfl_iE~Ȑ;JhS)98*%m);IVM22HHӒBys~% 4*B^̗.hgd/%E,$uA1^x}]YBV|ZdC23JZfڧ)YҧXLS3[(ݡ$8i k}soDnU8Q2Y(Jg DA,@N(JkIr4H:scNmHxNfLI-<NP(& e..ZJj*P@LC朰¦Y!'HҎ2g:%`L2 iiQ' "a=ubLwmY~Ŧz?,` ɗA=c$EKI'ӢƀcQdTխsϭeD)Ȃ>4Jt)*k_c37% >{(.՝`Ƀ+(I׆R ӌT"yˌBi. `-OXUAEE'{'0O4v/uh;n2Qh0P&ܪCZ]Gd.Zn kBhcnu+yXLĵd=(tFt-cwk.4˺À0N VXtؘuȆ9TUAP#{*IUJ':W%J) [ljU!RcfQ1LhHpu'8z? xH +BS5 7!:Ci*dL'BXp6LvҜ ^X szeY r3Jր1ugPP,Lh=4v#]7cEfJWukʃ AwŜN0tAC\<04KPI!0ά :P%@ `-. :2]:Xi65lD]`#)#A{Rl eB+mՕHދֈj \VKt_`3/oHH&9%R,;%Eࠄ`dV."2Db0T{4 {xWP>|p.#hn!:38ooA5{^̈KQEn"棊1͋BIJBmUa6rqi-kNLܨ]-^u mz`-$ >:%@uP< *YiU2Mt%C RbtLECNZ,J茸pΠZbM=A!%DB>hw6"T *BD :$pYwQ@R@"Pc*2Ex(%%l-ѧ@hg]+y r.!z-bB bjwXC-D{Phň޲A+}Ƣ Aa8J"Z!k,B$ fj2RA ٕ P?AjDPqkUEAeXYvnjpM"β] [%VmZbn+i=li"Tf%ݢZMnU{9 i AIX@iT_tВ&ƐZJmm˻zvW7͋%@~ _i}vuluT$KgV P]!lJ''f= ]Z`[ DiVۆnMEѽT$I"e56P.zfh' <%2`Crh[SP.O(7"fh}qnQD{}I,WR SAA ,3 R4Č,mAz |"2PAzƮXB{߁yE!8ij)ڥA$\s; 9,RAy/aèIP(EeQAR܌EEH,{awg=u @ +IUE :ƀ6mӺ`fav+5iQc4A&EJ3i^& &ci@Zv!; tsrw:-zSPxZ*;k*އY[Ttn.  :jNàe+J ֚'zʤ辐X?Д n F;k|T(Z{*(=uGm(6tqt :^#@.* w6\4*w^. b!C1 PVR&T$nBq$@Br6B!(L ΃rt 3 ՠEǽy f[6=I/dSN 5 rVoQ曭#=:M;LG{#ԇqU_l7X|yy')6-^C-7itqA;lrzqפ]t0vys~t3Fo5YnpW77W/Yh~AV\]b}yLOVij[7 ćжzv!]^,{m $}&uSEZ3%όǻ*J끜@a@w"j1{'@RP @b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; Nui-q$'{;Hj0N 0w3 A@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; tN ė#9,X([1Ⱥ@ - tN /b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v`rђ DsZ33Fv(D$Cv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; z{kW.N~C@Kkuu~Q7O0 ɸ9-:ˆ(h#z6.qCO  n"qtV#+TʉJ])+(tEh?O+B ҕm74Z}J5/ZՒ. 5EK*Λ e7bj u9vWxB9x?q ruWpS<1V?{cV&Kh n.ɼlӆjyF}xiS٘^~h@}=}J-y lz}~u|=Ac>͗>{#T3b3O9Q'|c2{|È>5a?9p-/OG@gJVNukˁGnP~T~p0+~[ pi۠pj1.>B;m1x#dJ] J!NWRY#f .Jk+5mPCW> bp/!g}a^hÅBU܃"^[+gֻ0w` ҕ (@tPt6BW@lPntuG:z3>]mh|y妥/OKDVW?߼Z|qVwT^3;Ӳ+z|h{{ٵeY} "kwY6UUdhSMyӓJ o/Oqe^[F'[GNHH|lϜ(*F=ڭiSn͏M"|gWZvgA} W馼z'KpMD>[r{^?_b}w+ 2Ο.?_^lݛ~eY'[`8G޲_>~{\_W^Ϸjn"lj3Ju^r^^lHjusxg o7 Cuُm҇ϯ&^~bm;ll-#\ ~x,WTL}z?Wwm#I{vW`v76; k#Qv[wj)Ed đfYUqFGτV()5έȜ Nmhx3$P᷿ݢXװo‘5q1zGX-e>{lYfGP7n8 .jdP_WذvRذ@UIUKJk '^pbRp(u^(Kנe-5:t}L@EvΗ4L<(Q iBk ר|^4H&"1Y<y& @&Bb1J&i01@2&Q hMIF861Ѿ09[pw⫽,{UB[ȏ|m .DX$ϻEu鸩6>ZW&; ~!'~a-`H}~~ԲՆ5sO7Z.?(ڠh)KI&)' I (#I}', UP#'Pb"iO!ZGPHeiA:BZ}ς95ϛTai-#ge.*|a|]ATy7sx_V6)珳rܥJjz}awEZvv͵=ecVc^V~s$ fJPs1$!PQ4bP8XJxT"XQjBHΑZĀ^{-TH`c)qb &$o95cX7]g  tYuᚢ<}F׻IZ^݀Ew[=V~1'٠Iq2߸cBYnbR"i'e&ELAjlY8 سe4M&Iж#$s#x{ؽsXcrZ& ؾR6 Z{@8JfxN\s" Rh-1$v1{wW"$\p Dъ,*5!/iDQ"pD_XģQ=9k|8Ĭ/E#jDֳFF4^ˍ'(b 4ϬdiるdHJ$^5@Tm(8)<(AN\1ֳF썜mWP/VE#\KNՋg8}^HiZc Q#A t>H Řxx)78U!?MAsnvU=VxWb|vhe VS=V[5W _(QjOtH`jRi,AIɹ/&?Lz4@J JNKC8!eBr@ȓfL'r@Hf=ED>@M-L -ZG& :yxJ8PNіE ŴTqxL &q*P#4Z"@'DLȺ7r ;ټxAQT%ه5׫C`X絘3_S&c:awWبFE#A1D9hx$ !0P$A@)]O\>/=ԁk?Q|D/#Rg;a')NW(s8lH> JwH4҂3 /Ԋh(bH>6YR:߿z0pNg-F-\ET' >ipBi7l7pNq 'wMkmg) A*Fn9FVhT$iHL#JF~GP}Yxu)i?S; Tr蹌DS/HL!:d6A"2OX >qJtL7Y#kNqs5nt7UDp㹐փQB%&5t- /TG8 CvɎJ !herw)\Ƹ]6ܯm<K]BDsП1/'-vQ"I%t6=|,w/b8:h~B0*5l+' }Ƭ( \AW_]=/{S !Pzy~>[X⇘,W˸!} $5LGF By Y$yQ]DX^%dP~XZ`̣ R:Yt8zŹ/cuuT֑]WMȽU%n]l6kw{QܦcUkv:[غn],[sZt٬;\m4z|&F 0/=x1<ԑt~Bתp[ޓ"fWMMؿk֦/tM"k?i{!OѯU5~GS˼ ; ©(OZw;ߍ ?*\ZlSͭ=8?\_g˜.buRIBR+,LXOX_AB~-VՙW]^VsK*#k#WĘpw2ZdLzB!DɢPwSܠvkە_CH .vVlq؟MQvW6xmUs)ocy;Ꝥ:iK"+@ikd\ S`vP<(r T=4N@ag$\ihM 6Ȥ_>2$HVABTPKq\< "s-TȵKb )n0m4*j30H}oR&/Fka݄2{,8d0tg-;L*x"{K}(,Mϻ M7E-8ځВ$Ez,pvR:=iIIv?%tb=e;OAO+Q;Ԧ^-> nf|rJdـTic79妕}jCG]4`:NgO9@ǣ*. ftt 2OZ649hhz~w_xQUr6vR/o>Ğ7Gf9Nz \?=&i~(iԕ߈Z](c!-ʮDArؖ\Jq?-[ yr<2ꖣM=M!1^B# `Zw LJ\?72_{D2\ˀ;;Zd9zWqA*?ajV\ W81]iYj\n_ZYng@wlT(v_Gx٣ .bmCJfoF*{M(u,ws 5A~d5)I$G|Y(Sq}NkYkjs8ϣ!:&;涺SY~9 r|̘/{&Wg)ToLBc%@Bֱ濜y+6m7 LRJ+阠x [wmYA2T-٤~>+ټӯ+ D;zw~CG~e/"40k(Gc_.6#wuW&F?e^SLJ~-߷zHIԃ4bK3CvtSU^$L2G7hq{F(* ?&j8'{U$^TQK[k-<ʚY &0Oh9S`CN9 Uc" Z(|Ύ_\^!T{^fzQmW^-½Ǫ}>?}Gf+R]8?;M VY;Wr:Y*ä2S_Cwkw]dh-Y?orW2d6(-2Qp*1D2:INntYrVgys_nrR+wq~VNHnq1Hd1rNL.Y%ZBJ֥Ȳ^gmmFP*i9 cH5ϵLMZnI]-! %kRM??}_N:uS:((Qm], ԨξYkVŨnbwW'mY%|[,/D΄z*xQ2CK[aVԫi  }H}k hcGk\bJ-\ #t^3.%<I|: ھtWFh;f7&aurnL+/];Fs$;FXF 't9`KxQyՋ_7y{U4 ,n]_Fl~%\f ן~LZŶ+*n&6fw:'%ShvJtY,3^_MF_7wSJjJ&zZI ҽ4t2o_l`i^|7tb;-C56LuTUb/?upq nמּ]r\Yez͆ŸWє,,S-:ll =s}Gdz͆^qr漚~׫fNoaP$z %f4K f[<0OvX|Ζ\kɲY1[[kN(>Xnl=`T 2©̙&{J )e,K%I!;1lC`iWB[B,nR2Ix3$F9֖8wap1%j@gEwV"I?POcg3tS c-E(˭W+Ϩ&2A4&YeNAhYPvɧiWWAHåִ,&EF 76 )L*&.&[Gm!G"+sOu OžrF *XL!*,m 45R[-H%r)J2 μ%ʻ4&50CB5U u8F2sRdtV5ntv.[2vӝ[h\[QcjO: XPxJ9+*&.rL:gΓ9=S(y: ]{AV쎳Vپ^hl;@w/;|3/זL*!P\ْhR*7VkE-:`=Vlk,qɄKyPAA$ȉ*b!n)$yp`xshИiݙmŰ,}Eaf[19lYō(xF_s.dh촉Nf}YVk-E ~:Ƣ b%H!)!hTk{>6q:Ilw*JN8il<Pd,m4Q%49Th*nd?w*F ( 3YIg@KJNɲhqp{CH뤎 F!KbH"*r%<)$2/5,4DG/H̛SGr)Jy F~N~wϹ|+?Al\ \?(^]N7N9!IS`J&EYfwjK&WIueȊ 4g>F*=<C@q0zLh#M&M OV{b!ͬK%fPinhv)nJ/Ó.Lԟ+E\o>ޙ/WߦW#ӳ*;%hKrp5%ek֊7>M5Z~ko|q~zRܗœ/aO7Vw 6⼎ /wh^ʖPff,Nj5uJ>^iΡ+[Vg\ꪾJ͛`ܺI9f:R4}>eJQOK8ި97p ʗ{˗*4_6{~w?͏߿Oyo7^wo|O @ *,n,o ;Maެia>M6{=}i״^Wjynm|mH?{?_|~O=Yzףl\d *p=8ȴ`"ݟRTΥT4bCGZ*%f&&d'I"I&-xul2^U}J霴Uub0֞otrSt6@$ [Dv>@ >!$ziG朕cʾ^hmNYXe*֍dwi\vzóLН݄T =o5tg6(W/Tz]<# 1WS!2Yd\Ȥ. J$x\+ѳZnܳ-7{A4_;/2J}%:F C,{I7vyR0^>lb 4r)R3QĀ!8o|<5q)e?id\.EmU|?Y|04f%63\;v#v$r}v?;/gRb8-P`Uō:VKzzyd{̣(H#Q"LV}JXr ffeĔIb Y\Qy#i6zq'ef쾊BTy͠]ϷS¤nXz.8|~G,E>;>"֨O컓!bbsPI`pΫ/l6B^UBO u1W`1pD1MBf[IΣ4Z'r{&B~uyzh4}8}]xk75 >O9=Wϴ"TLz˟\ }ЪopVW[\d7O%O?p8[.>כϯ]fX 풯b11r|^ni- V>Vve2&'*/0deVDȘJF2 9 qY kf1˂׿ç2UIDm)LҎl V`MbQ 9̹FYNq!uۊN5ޙ~y_| b̛mgȾL=5dVaצx SI~cL<2aU*'Š:ՙ2[;ػGuEpX1@M@,+9snlR*%u:fiR&ch(츍`L0dl๐Yء>~)+9KՖbk#β="!yFs!QPօr($xl?~T$¹qǹesYT|Fb:jj `jq ;Z|$jk)-tJ->dz$dfh5\=P`Y|'p@) \Upojj5WE`H\WEmUR!\q`J+X y0pUĕPHU҈!\ R+l۽"\UvA:zpW$x0pUUPHkԾUravp|JYLpEk8W{ *nعڟ#\)B])ES{oяcSuN#zQM4G-Ǵ.h^ h 2EG5>o/w&="gQ$k)F:C*6 Ҭ#u:RըfeGNuxc^^6煡fFRO_ΛG#܀&3 UiT7\ $0 - ~(")%\ еШp-5W$alH gWkWE`u8a1E\s07$-?\){pecGk H\W$}fR.,s  1\=H\+wiՃ t/ WպCO`g$$bx0pUq(pUjH)gW p@pŅٻ6W}Ibȇ$X`f0/A7lorGdmI.KVan6YO"ɇl+D "yRtGf:BFF+h;+B{ 8gz2D0+`CWW).tEh5NWҋLWGHW֠s% ]CdC|te5ЕމdOg@Wd4&a4>x8MLh!NӈҊU#M{+U`ש#zt(v|rv׳ -$`3]!]o  lj6[ $OWfNWX8* >i82N*TǪ^:-aDWX dž \hZg$J3]!])ՌC`CW7x.thվk;+BD#+gDW*6tEp`:]ʐc+HqN`vEW;qhR+DI?g:>^ki6OUZ|0H(!HW4G٤E cL-̈Nkgfՙ43sjI+7SK̓S24,EiYwg-'>(J32;k6yp1^A$\jҿ4c@iayRr{dZ@^ٸ! U6.0$m5S| "O:^p+BݶTy @OГ{WtEpBWV'L(m޸9F `DWX ˆbs,Кk$+Sz,ku`0 7yc11hAUJ#JgzeKFt{=t8.y"2]!]Ѩ F+k%"ΦNWgJ/fDWXi> ]}7ޞeЙLd;#"m>u"g:BrB';"wpd]Z<]J+ &@3r$3p/D6O/;N; I\g4=TW(!iRA`)c>WBWVC>%!o0#])5l jÆQo0 JBgCWXH EqhCi`0te2]=VRz]`}J!u"yW #B^6tEpBWv$+BZd'+r]!8+"Jq+B{wq( d:B2h%v..tEhC 42{WHWV#"]\&$ S+BALWOBW4 CH.?.3g{Vg 6̴31BQ&;Q[6YFM"W7^)N}(uZp:Atʹ"gꥐc56VtEbY@A9Cp-}B?T";̂R[Ɉ ]\wPLWHWup 4fCWW9sLhū29J%2 gTEp=08ѕz;:|*?w=_$FЕtXK# #BV+6tEpBWPLWHWt)3QA6tEp?< m3wtE(;ug&p UfCW&]J'2]!]+L;Q"p+B]JZmvO0HpBW+DEc+'*,U4 l:!u{odSvT!4 M}ޞ휐4}<4iڑfDW[>;׳%!-D B+pR&04pCJ tTyG*8,Q"M0"yPZtlT (A<9pK8@PĂAAW.cU/] 4Ή"j|#+ѯv ;m5 ]ZR+B!ҕ,`7Ή֊P:ʸ` 'Bm *Ʌ|š*ҕjM*R2rQ0^f8Xt~5,.OnW]]rF7h=(ɃGx%=l74݀B-:xvY]W9u(?sDzy}V"n-堙MGt֞v.QY6D h6^%.7No#`Bn볫Q5-geqUinU @H//f}Y;sAл =6Adٕ|zt=OW/'C9~oavo7 \k|mܾ˫XTa!fw Z *RIKN/Nf|^?__\pB)U .pc[=ȇBRAw9xHjg#|g/G`JЂ.(є:DmJ֬A{T)=ȕ[Q!laJ =bqMFNlzT\#oB3l3A!^ kAp& b[mUwr-Z[ۤtVY=;Q%-J$wcRd'Y`%9)Jq-6ƒHa9QrIl0F fw?b49˭ɳܸL=Ҭ7YE)9Hql$q(zFRUmm%7؝f޹Iw) cE7hg:8=N'mtrzw9gO@׸-kZh\r'IQp~!ts\-P59}j>wqk`sDt\}Wn z܂V=~ߘ+&4 \5&Wɜw~U p1f[*]Tf@Nd߲ uuyMTzv|yHG"ҭ>=FdZofCtFtt}YI^_ΦE5?|)KIYhɠ5:meQӀ6~24M_޻-!@}!-xttJ*.m?=޿iOtk}tT%@thopGFa{4pϋˈU)vƆWcoi6%Z>o>_V?Mvj|~O9K:~ dmOϪ|Fmjq1-Ok?1\T=mDl W A);;l3}ڠ??ԎV;_۫!TꛝRҤ^ҠAT^Z>|H' x?6 S? 乙.~\`M/qlbK6<> l sK~ 9Y@o& ny>:C2N=ȕ[JPrd™+T묞OnF!D a7w<B`JpvAI[CVFg7'oiûkCzwYM]WU‰4斀ʂsdl_Tޅ ɆæhQǤ;ķK??xlBt",[Ws7Ԑm sEO׷˄MqU_RdfQ!D^D8YJP@8B):|mx4BJxQI R࡚J BV[{D(`BGKEZ%^ =;¸s)Sqf騝L-tmm&]fgyTwSKB/cX鴛̧b^E':%:Ş ~K+raEE yQfjoҤ=HN٦Q!& @=$d&gU=WNW4_h@FiοI{1i-xQf⒪Xmsc*x~8 Lقus3c) #WX9qڸK/ B=hV h(.ˊf:>tfMOPfqȀ˞n\F}{ݔsѳmhQIaJT"*ogOo3ïFQ *XpllYTUn*L# hUNTq]_'[(C^QLQQZv2%$-lӅ~:w~RჹpgqzizV?#' t~lMf`>`ݙϫ۶:u?c3}aĺtոR>JԣoQi U/þ~e'+RryMng!m8k +$OX&سRa7q4˷¥(viʶ(r9a1Q>Gl|֞MvQeٔ7ٞ/ Wo(rcl ߪqtz(3U agڤ{_٦g =4c^d?@Yx DcɪZ+A]cZygM\6q=Vg+4y7YYxrH2Zyxߖ>f~at!L|Hi%b,U&pT5,Jpq "=oH\h578iй9l2 tdU l"8ďeA$#=VN7ٳǶvբnoуb6'؊O'7yB'_"| Զr5.׌&uhԲb$xKl/WccANiϋ_КحeaPM ui\-@VB)꺞'%u/Y֥uU6F hv_mAיzR1]>IGxszo6¶_[=1'W9i@2^X\QJ(DM$¨i*DJ6nT0C jlZ*Pչ_D]5kj}ZhTczAՀ]_.R|"(%LO6XH#Ntz&e QRg=ZXf(ش1kD ]t=tQmLfzi[fE^|f J _Vq:UKhe6xKzZg7e\^l'mNTxNI`sv {1HJ4ZUM#'.pF0FTDPBY \aV6 ;Z[=B\ӓpjHLCZHͯ쵼_/•C)7s7kC;JvE? cm1 ȉuj^>5)rOo?jGu1DqqB "o17H ē 7|\Z/J.g5ז t%te*çj>dA(%C3zj9ut/=D}Xm|YUgRsRx(zn a\_>WɭutٻƑ$W;@=0f{z0 ,[%roR-J%Ry(6$_v}[i{Hid=EKv2C#-Ϛ;}u۱_Jޔk:nS|5]wK'>Լ(@VзVl-htjy3HCD~hBdV'W0yy'\<,QS90X2KxJK,IY +R:,B;БnjTfܓ-~e;])דMIu&!,Qkv>TLSsTd}E{|!\^ Mhym*NDYdcsM'C髌Q̉O\t„<_C՘xRnU&ubF妜1o}]u5칚' J=Ŗ_Yj}kA } ȍUG&;m'sK_ Scjҿ6$ߍے+cΧP~$(VM1@E״$J; f:*k]L~wVT׿~(f\Iۺ% Hg ~OXl:GBjn@飚3߆՝CO@v3g|I]HJ,=I C OPͧe}4s+/U)2UPhJodҰ$-\:1fy@ƁAE/l?] ]},`/]6`Q֧sEHWJD }A^!>)n֖uʼnHT/x>V gx3ZwB8=1s6Th`\)8@E eآpiѕ@: SP*^DA`)R6˧U8P>Μ9ͫ8評]U9ʥsY{8-CbBkM$DCV^鞾ͩ_S[B :j줱4W,!b@܂Oi𤤊tl-B|%sy*i*4Ir,#R`+bЬ:6NRH8/(+qDꦃj'7ꋰ~0ڪ Tב(+>3 AOss@V]&UBgR$K"G:=~ YRYHr\pX`݅r؁a&,/t9Ml@ +@0Y '4|GyiͮJ6HF%8š +}%ԲI[V qnzx,Z-$ =''3"0U1H(!ںJ7b &X8* G. qo{G|0\ rFS~:Α;$ D;jWAF Ƃq`F }A0EM~D \RgeƝ`+us5W_ Pu 2Hl.MPBnDZ<9h\ԍ\AphqhɆ *w|0eLbs:(3>fE&Y:_A JTwPnkj]\՝%}܉-ĉ<$_)"P;{`7q薆e)uAr~v$nTsJ];B ]:좗J6LPj=QB78?/qJHBeYUS6gEPs92tV:sUn kSCՂv׊`Q: bј$E1@™#x_`dD?\D+>7;ѾzNE/ƌ`]0=mʄJJ nj;EOzP1@ꮀzS5*RxU:}@iPXMn,0z{M(5 !;'v@iz_uhU;{>hPn`6rN''$s6, Rӧ JyW}( pౌ#%1=fTs {Cx yx6 %TiR&^˗*.Lj,7})~C?-dahEXE.y%_jNn\/'P;܇#Y!+MO#Y ^}\!pӒ3v^<./ÔEޖD7k IgK7-$l} +%|fë"D:a`?o<}ڬD`[`_ `NL nF%X e&z*F'͌es3Tt3&D 7,d4ZHNbXQ)7U`_?G e&iZōtr)嗏ZKZpz?R :Y)\XyOpI"oUl(wx Lj3IR/'3Չ-9:8bӹŸW'PxAj,Y;~iFQb?fC&Me*uފiL'p;>OUgx<-%{s?,+fNG/~? I/7wc~NNܝvȄAV(U),ˊ42MA8mP=<ӫ*Ż%DV\b+MT+CQXΙ&4Ym?I5</@1涮!^] р ?De: u37%T?[o6V 3rFbD^pb8F8j)ܭnl󾽻1/!5Z_k$\[pmT0i1X0?qS(+it4xviR$r,j Jm}s}/r%=w 4($Hj/J 2y<"vfm^3Qc]kANVD j :hhFi->$:=jUun% \hZvh)Į3''wǶ\p~ bVϽ){%#:)Ve0JBrk[[i1ͨ95> ڝ峝foJ穘:OΉF>xj1PATbUTX^[C 1ЊS=hjA+3`=L+~Jº/k}3MoΌT9TD7y%BZshՐ3Ϲ/O)QK|yR<*ˌ] %pD"-X^dQ;]'()5YjRדm^zmKݐ5UV}<ցRgӼolxǭyc~>s֡ U?aAc}=c}.ʶ5nɶ1E^Ud+q0Jb7 YwFA2{r'zZ'zn ] m˳?w1Lc[%@g8OJ-{T Bo JEX@f9X|eboh "h~֝:h:Y8`J˗0歿YaH ƯNQ,nEYA6NGℒԩ|Z P&bҔc 9p*=$6<0NIsӐu|Zz1v{da=RU =_;-/Hwv{#v?@".R32sɑK5g[ֵ; Δ䗒@z $Wo?JA1.k-kI]eŃ2Nlg..'i-haI-@ɢ'b5mW<3KyQ{jnq 8O'd$ցegcRnʭ%zbr(Y ~xED(I`8\|ԏ8.c%0 NkA~QRB%D(Qu fộ@J1jm(;+Z] LˑV .!6<_TMCEJZawS%vW^Մ\^ ϵb*k@uHl}A("q3Qn%u vɓ4&Z =&wTxAvT; -(2 }m׹EZĝQkG#X2Kս}U ;h}`b]uKl:avW`o~C^aą"*Z(Z82hv(7x"pPk Op$}CzMՅVv`/JQfnE['fcVo_iףᷤj*y: -4q֟A>YKgd`Ayų16HK-1 C4PgYud09돜z>3\}-$ҷԎ`{2=34XXo@| UpXXlgk_?FSˆƲv3#]_sɲFr!@=Wd%PQhxq"n&;=Ȕˑt"s>D,%w_q"6rL DF\-i7ɆXоXY=3X ^K'V,(":+~yfuU%x0O3PlOUvM;>CThfppJ J~rSyߊscbs38A~J51ifsiJ'ܿ:$9@Q2EZec=T?{O8_!nj}E>2OU$muYd)")FmnGA|x(t%7b~|%ܳ>OjSq$䢰.[wQ\+ڡ-40mD 圐vi uhSݎݪDe %49QD1eֳBѺ1ߡq]TsZ|!a*IʃW:oo:0"}51(nxM}7 #r4EBI:2I~yqk!z1~^uy SN+eC)*[f '(cF; $|k^G9} WZl`yy}3Id=Gv 5&F#,]R QHMw!4'JyY}I=IS0'@6+E܍0m);RaUoڠ`#3BM$I8;:MQ:n @b'R+WLkx(#(\#؂ sDEY+1Gxӛcbk2߂JCҳ REqq5YO_j#ѣj{έwD>'^ ;V+6N Ŕ<(v6m$qA\ՠkAP+-.>KgJh\z~U/Fi O%rN:?8#gt/AYs*\{1/84Sa?76G\0MF?,UgΏ8{:OkdE ^Se~K u2vb@\S64T~4v#:`}~ o{׆~. A@a(9 P  ER*<k(o+"(_3[aOS X>ůRpe(t4GU FԬ=lo'SmAYD7)z FG #H􍭿 ]w`hX3nU-(yJüc,)fZTdD #łJ4xyQR] Ux>RW o{e[LlP28NxOc851M "$% HDlA&ZIgtQ'hE|^瞟HkB)CbR1p* Q/(o"git9]]l[`9tݍ\o|{;IF|R8Pt'?fLqƉ*M]A8RoS1F\޻t~P 02.rCOM*2Q_U T{Fv~:qRj ߟi9# `;c _X)OSlV\ ƹK+&m3&R̊QW7O/U9yٷ$$Ũj=Jk!Y$Γ_ʴ]&s8-, cYLʫ~T^%,z=뵪,o`|S7&$ Z^SGK EKk0DsJ5a`M7 6Sa A`$z~[6h^p4L6zy>h{u%73Ewpr;ӽGcޝHp=0X%qSCE4pnŖFm=Nm?4yŕW)|{<@+;Ŗχp BBYX+4k~ 8+6ȤjrJ/؉QhM ~%_vlv8Q}#o\؎(̯em;`&$qϥ+:zc]>_Q6Z Ie쥉sB<}<ӑ[Vq[i?cuӣ_tG?Yq_W#QO=xqB1SD1ß(PLf&?)''7_&>IۧMx.|&}k'2$IL)/?-G?/FpRIV}J|V/GzD,MV1Д̊dE]p?.,DakAiaߧ3(-)N˝&On@1e-1.j1X:q?E!Ŝ`GH ~O&nkP 50ejl laX1]^)f2K3P{*@W xbu՞e Zv!,V8t7X,8DdB-&椥R {&ϝ@Xq*s٭)@:)o5@5+լ=_]}취4Bs-CBZf6)X||mphe )T-'(V~q㢧VqN倮km'0"tRrHE}+Ȅrm=MMJAww-kAF=,= 'VodL2{);Ra ^֝0!u_t~CS9wԞo+_[+jiO?-iVcCF)A9!qV(UӘk1wĽY#we+f$vͽކ`Y2(\ij"ݘ$P!#-0m{V*wrU'^:飋iZld3JҵUchqe\ xǩd' ^&*(]65>jgZjzW*os=icCۙr-BҴҽ|6 ao}':n>{P{0< nu^QVM/~>G\| RrP}kヲh;O/}f0}V]\.QSȏ]i?%ȩ:ZX5b'(R\K)U Ʃ4nqBg^)}tl[],^[J^;^ ;ij=gF^Խ\zf^\qe{p19mA;1Ʀw`PmĬ):ٛA |Poc3%᳚/[އW1cWWT j {{X2]W"{$]o-(P>'0pj^>\ '.S[>Ѝq`Q7,I CRu?H{ק=H]%?>~:qR* 柖cOj/q>jS1Bd]p Ȳ9*~[f(bt@0-v-"ÒpjN'yj~?WbTN|^/EñΓ_ʴ]&~շp$$tDzY0W(vqUB'&`2o j_תҀ+%X{6f@l:U>EFFeBY^$-X^/uCaŘyAhjUSn!XUSCËrVN+mxM|b]a>n8l ײ; ro!@It\1a qs|MA/|"^H ckB܁bYe^ɀ.kTpJI#֊eR3E(P> ע?Z+t*C7/pB9b lVeN̢4%s )EOY=^2^h9_D`mzlxEl\.܅( ^ːQŔ0'"<^Xt-] h>@>6m ;\u[@!9% H;@?{۸Šu4$ ,lb& Җ%`e٦.ċ$$CǪ:2mR$ i·iNK-+\mdZ⛙QutM7 X$P[p@o.eKK<icBTDDM3T@zu&qAIě턀h2̇_MEfLTU6-B4doNhR R t3Lros '(7|s:+A d◢[-:DOWQ`m-8TIB M;w |=KQ)LmV}~OMT[t 9\Ktte1*7RJzD:ԟxJ##%1RpX~q&wϸZ*VMQm%8ƍDkg rLh'QYh$`[2D{caj1 /"%hXjYG9^(/wdw߮%&֗Jtt{ ޢ毠)rՅؒa*iZ+'[arFkȒ4Az4 %[DB!) tznBtWM,2Lޅis3kfVwf@Xy6?b((BpeSB5뒯x.p6K,ȏIAʔgBv9!cS&[@n5vNkB +9$st^ʒ׻@d:QVLY:fcw 7wMI_-L{dc2c5^Z醋b+Popjb j/"dT(Lf|JBX=%6dޑp VwmV\Ť<U3.擱ޫ)]L5f6Ta Cb!SB $f7)v<]̗YFwl$v0ʈ q{9m y&\Y:@{n*bqoP%9S۩B*ǙT7WH gzJszɖbn1!8`s l<»FnjX=&=~ {;]Ndt*?ϋ:$5Up.WxWvMGP ` dԂ4m=J(9+ߪ?rb'gEG˺^RS V_fGuofry3#_,$ MJ35(t?ѪqS0^8M0]};ieSI;AD)R%>XېJR@LeD>X GU XoXD*mP@m5MA1L1[eh{(*+Y'\K%dMbna>ST{f*cQƣ IĴ DGW/[WXl;gNhPa.wR.XY8aӫ )փwc6 dA]7A],]gbY߄/|"߷\טuSnliLN2HS͡3X+awTP"%D8W")=v q_;/$)I b(u²R(ǵT<ϸnNk(FO+KPͲ,;AcAhG)ՅVM]-4 ]׎ѿQQ8%E,!ÖCHoBPX }On q n0O`׳\&7􎱙}=]wr0f)JD|Hr'i#JDžܧPmm eU7QmРKx*DGI8j, eǐ^I2XVzc®Ǐ0$|b z) -Lt5[87\s`VxQsu 㸥:E|Q{Jtte3 wL;AA_TJ0UX5yUЀ7g6X٬H]擞^Œ'8j9*ALNͬKdJv 1ESFȻm$K&BbK Ig! _ ofuZǢn8r"ictCp&;QW4up׊AN1YTFLr BDG?m>3xCB%:dm_kB!c9k,LneݔF%:i0TfUPQNwnD` ~MN(:Fuϟʹ{:J%<=?{2m|ӚzMa 81j0_n$&YPRwE  k^^50[|KrsX~3lUsmξfkjNIndv:I.$̦sT[Z ˹7ӔqOEU \/'O{0+я6Gin[t*Ǘ!uDNzU1/pHQIl)*ޱYmyW`BFU?T͖cLx2s1xc?QƜHG~MS $y|}EzF >~Q?_y湊hp#d>ExG{y[C˔jo8?Y>/ԫ#83lĿhFog^OFM) k 56Zgrvd=5f֞cbP(V3Vd4 3xqcQ^T\(/ AIͶ֞n\ "_JGjd&X7kȟF1p}㽖mBS_iLvďN)U] @b+A6].oyXf!~*Ɂ"\É_(o-}|֍e\)DvlJ(kJQleBfbx9WP!@4%Sv* -IHW{/"K5@KtJ6',wf#qBƩh5$Tn'gu Kʯ? xr/֏Ihq&>gk!\~8c ˍ Ν5w^=]]lcStᲇ;wm2LxL^ɪf&}݈=!H-6?v>j 9Pi+HM44?NǕ7$1&5\nWÎZMjqXK(('yhTBSDz2f":+B%:: && d6cs}obEJPE ^wRR%2[m%;5-)hP]Nc4q7$08Ԯnnh/X#VhWW'Um{SN?xbw46 e~QN)Ck%(Q~#&D_V|' , |==AL_¿ a$`L>a=wԛ \\8u2p Ç#"SQS1R%xvU}GAW0xvDqhc=ۋ@u98(cwbuSToB} c R;ol0ZwZ[UA6$\=O1 lʹK@MP*⏜#;NHR_dh)YXSU?л/sK;4so!s(Az3P0 4m]W @Q, 8ae6jS@3-U7OPKAI/fqL|/?Vǥ?-V?὚OIS*%E?Vᅨ x[r@))/XGE9|U;_|FOIoG?F%o3qZ?zu~ZMw,I'?Ei:Lh V#4z9Bbg5$W$S A)\SɌj5_[3WӑaG98ެ,W#=Dj>/ٌZNfuI=?>TVu !ElDG;IiNU V vpA++6D hsޝb.!a9'r3Ьd"6IqD{i6'ny &2Q (*sss\KDv=GL/JgڔF & q"1uT[(PzI/4#n@!e' p EK= mo9@a7oƘ߆qArB)aFdPu%d|  W FֻTcmM1G՘5ԟ 5IeɶGy y57 \AxKj Vn؍"X Թ8Ў"CY"!hf k\6Xgً[Rj96S!mń{Dk-.DTSY%:`@2V,Q}Y=E=cLJL6@WAn?a~ Mt_J)5|5+4 >d?>;8;98jX+R8sBCό#AvAn7MiڊT 5~Α ~mx O(AbG8TIB M^w+rbI1'1*eI@BTu4TmmBIDNY@ɡ|EXRRy$@P)KlMBar8E`"UURδ-UMl :6bl!D[wl5"RJtt{gIbu4?㹉*igHLg cSo~X&9 12BR$+@ 4܄Rc>|ZgӉl2Wewґx\b>CʨUt!a0Z'ρ("Xp]- l7W όs{n42lT"OtX㢼N)ۮn7Gv 2Jt TQ?FQF=9:sRo$I6(i hu8Jhi'˕B&(7ׄl?&I^=cQ~f=ФLW(0$( Ӌe%*Ʈ(3>X/f93mDv9bN g=-),^Y_:o`T>Ew71X榝kNε{3 -LXP%Uf=h;ZrAjGe佦Xo&iL)O6ZtoB=SI_=uWW%=fZhߖ2TkA,.qۅS+ ZZ-w9ɚJC% Tƻ"u")~b[^z 'of~rKbYŚcKsVecB3E!|ahVO%}cm!h-5l{;u[̃? Ss.mKz_{?pZ ͝6fºrYTh\JVkU8Rir$ `#pJ9wMkx5_̀'mRMăT!O i|ZZ5$!{T|D m]O#M6u rh ?vyA0ca4@ e7?%6 1}m},/ MLثuhc%ߚ4gG- pAj7e֜{ԡOCN[){}.b)rVy7 o!t/>HB"â.wu48 ,/y_hwQEK˾dq##[t*|l WPiw㜾W]r'4eע߼$9&qU  E. GL0Hzv%$ȭE42 1dj*43!gBrҘ䔌Iޫ#ƆC%H߮jWOg SºA N-eE00)BNw|CR^릥*GRϗ WpZ_;I!7|2;bg̞wvS& n;kݭiN;;tzf 4I$%R1 '^lYpJЂ׿MGVdI@m}͓4b[ |;`Pq6 z>vH]yF .ҩH0 qr,#dI ołFQ\ιsD>myy!ٶgYeeO&G/%~ 9?+{:=Q-R|Z.&2EL ~-hG y``8C/j%KԆOAdT1}Irmz&3ѱ P'ʐbӤqG&rKpQH_[OAߔ>C ̌>¸e7$%',9(ac(hhV);.͌6+KJ*}ܐ50OFƹBxʒ| qa  D^oJ`)6ÃVS L9:>_y{TW;^=.<"ok:{{<@u 9D`P"l_#|[vO>ڹhM,K Τ\%eȭ'*hcEDy} t8J}b.IY [#dTGBY}`. 6KEYQ|m&&a`j)#:#"*<-f ؉\7{;5W)yD*yWxNI 6f cx E4;ˀp%6$%&$fC"DTեG6˖OP.\M2Yw@BJ!l\91<9WKkpDbRd#`#1XnGk tv J%K-#ƀfA')60oeL xj8gH*b9ٺ8z Z(Xj:2o9g4Uށ!Z0tQENjD$e8lSx+YݖWc"lBX2)?I 9ĢS2+BfIî^nN^* ؖ_OxJ,,sra9lmZ`dg<7K"/Ѕ;V8(qeLȦ\NΈ<`"EJ+X%L+* Or83 웸f7.J ;1 c:fŀڗn4Ya1[.wcL ;WY0c];G]N a ;cM=8; sW]\2%xa_F\{5!ZJx !gjvS?*EJ&YR7oyu)%KK}žtIN_\NUos$"΅ٍ;JȻ%$<FF8rBI~e>w.`!)IJib[x$pWs8G0PB3sg^p=W:ԴK<7p2o38jv&*GkƣKýR'>հiUeZ N ) eK6V[" u%[k6bQ?1$a2H5V29zY RPT;ꈁJ$pZlc$ƒ[m:p8&}x8^KBД x 30wBJ-?|@>%j?~hه |A&vZysm.ci @gp—.L%8owͻX|~~|@>= JXa|a>:{/8$ۚh/ PfO(mUdRk 7?5N5LH=df&aCk-?1EӹZh)]:} `#Kx)4=Bk䄰fpRTmh̪>M :LxK9 +̿ JQص\J?&{%3_X~.j)/+ϦD% 7?@OP>x~>_H2b99C`gbΒm\ %N&gA$ړKvIm~s-Jھ>8UWBOuLgkx4T &k^Ӭğ)Ţ}{C6hj"oON6(.jHBc{Qs. uQhf?&*>Z(p`>d#(F\ZhU6N/}@G@],[4nޏ/[ P܌\qa7'h nX=v?4J lb.7H/Bcw'}sCJ/l&hv=ߠs;<}OkpxpQ鋴,BFʍyXUJN3ߏB34:ӣ "%S[hEFШ|ji@t|ELAW>K4hdAy.u=7hƲv2/ ˷%};tr=f4`\-nՅjhMX\dg_Gt{-[3Um|6 p.%eADNWi5ɿZ3gYZQ ޻sҾZkEZPj}+>@0y5v0JA46똣ZyTI0Zl 1;[3JLMÈ+IVg$ vvN+v2_iaNBFfDn7AzoZ>P 'E3Za$VC7$ҳzh搤bHXVsN:{hCe6.pvWd~=$ 0jddaBb ʬW"я@朰q9I֘kutpI8VJh郩dg{|x??&G,ô+[SPvdNP>)2uoyLI6ij:OOOi<G#q;@ xZ|xVC >j̬#ys.ui$jwltlH.uXg!M*E{Myad)\RqyJkŋ+N8#K犖068GDu&R)4F4a*$E"H@D+ƃҦkiD6NJZuD r@%a^ EI~ȵO.iQR9$+c5|0,e2x%Ā&$N5Z$c $b^>ʥc3~~5X`* Ѱ?["?9a^aKzut[=sڞNoy7ԀQ`D9XnL>grlyf-q`N- g747M^osO$~tu|$Li|/z}k9} 㴾{C̱n){F'd w-`?IRG w{=ioD{}F\Vg;^!|S+qau,HMoy7Ap!9`˩G.t  [,+7R;S~pg q *걔vqѴʑ|xL)UUYW/]՜&X\ˮ+Y~$ o򈔛>ۄvrMXQq?1*":k}@hSnh=g b@ggKU6i. ;l=ɖW/^ފiJԺs, t*L׮M:Qk6+Z[!'YrK7uBhާR[4b@ZlhPZS7D) #EH4\6)-UOz SV7ytnCnf neY{ǭ-dѪƍZSwtlT6L[+sy$3msq2fϲ: Ukqotgkߤc+b/;&Zi\dH&S|~u_'Vc&_D@GAZ_v3?dCj?:?/O󰻧|r?͙5]-';^wQԪv^3=."6uBЌN88/9 y=5YUΫs\ZvGrfC_7?:N;N}TSN] k{th&P5 ".>]bbх1 yz]_ԩvחo=57qwS|}mLj]FTjI>hXW㵱W]OLN= MTpq;uEX;-Y9#TL؝ʛ.] aGө}oNzX*N\`@#mz{`oo[a~Tfv^8 b vjI;\La_?٥PsWMOAF~T8<ލ˴fh>vs@&_C"h?;lkcECMܠi]4q2OWԾz0-``JA-@XZ$#- [5#?Ҡ%DH ͡a_S7%z7>^/4hJf?_=tq5\e!,%~pũu[zodEiBp,./{K~wG7 sXʨSM$>gҗUc=KJ`aSha PJfËx[:8Ajݰ99-TsYN1Ӝl>07.*ɍ/%PX#2Lf QD@m4 U"N/~K"g+mvhq*\YkXvD|aΑ::%_"PPȐ8/!JFr$YGw].jT뷭mkF4l[#N=&,m:)뷭mU/#x$T#TdcRH#Ą*S!T3&\J"` t<,v $PB1 ] h8 L)ȕt%= vR6|"e l@@O]-B%)IU e$XTA*vΩʆo&OU2Eini=nr5¹-&Z~Eq3A<:,hv^8H#HZ :3 6̴kCǚ쭐ݕkֱ%mHp>ݭe_ĉؑn_,ɣyfzr*3z$p c9[!b^M#6Pl:Dq`cг,|>Q;6$'UWׁ]qMg0=c*8mER7w {Nl͞ӘX 1ƔdpF5~s2nAY&ғ~ͩO(%)KVP3t/ZO_}cJohS02kI~v!%m2yjwrȊ;LEcNbe_ӅGŅ/Tc#lF"#ړx<[LOdZ)brcf#l@XYb-Kf:q3[f͛rk7gӇm_G)9>}wYp@19C6Hg׻4Ѩ?JyQ3\.NzIKEDXG"=FH`2GkKUK>EʩSSxFɦvIÒ='CXx`&љm tS)bddCM5_gH8d),nKHɆ B #7{/~Yx~Y =7J5gMCQ3袢6xp,Ci&@G+ :e.%҄^A'3g0QȍPbF?H UcNUyHCGn.Y]Q#d# ש'=sc)N;S1v-Y[$roF!6ќ}B\HuJq}ON?u\n}?>Įe+)q n"mtU4wlWFn?B|^~üM]m:z}+Ww}9"#'2_z$9:7x,8һyiز ь}J[ޏ6c&^D:/0aq -^LXB#̈́#ڱ{%w0f?/3Ò5W1QݮnwM^̾[4E'A~dE- ܾ, #$93V8os%'9Qcg gz6~oFtz.QT 06L"0ny̥FQ7Q\^6M=&# t6@S:g gi,M"fΝϻ1YgI|6y9M98cY31o; g9.aX@4xr~j(ҩCt2,'އfMpU7sPatfV1ªY!z}0bVj,vFbo*vm8$=+[ld5\K?)%iY{ZTyG`f`-q\ zvxJőc.H/NɭhG3U}ȆQqٽ^9{WWyʣLKp<}ûԕ>j!׾Wz*nV ȏ򕹮uξE ѝlT -n W|?6 ??Yxo/ǔ#nwztawx7!(o6D~ޔqdwAF= _^ϫ{Uo{|Fd_ްaTc+(`7TxeB`o @VBL fJȯ0a9f")B\ JJvjݷDzyV3󺕼RL_dq\]`Cf⮙<;D\Jf[p>+v6];Q7稝Z̉S'a@JLyJAZG!DLH%A -ZБ0c 9d(1F=Ilz 1XPT&J!"kΟ}  bNgy+u7h2BS'Ly2w4yt!Ts]ͤz1Tc7tɨq$Q+p@ /࣫M@7sx|w` (HrKZ5VvB*|qOB~xZ<{o{~E<4NH6W:I1B%JJ!Ӥw 3_FJdXälCIwYGFX _,غ7@+U[m9;[wj皠g]\ƆJQ틐wLڐ//!դjvBy(Znrk̗HfI『t"5<y<_pxG%A{x<fILl8dcwqJ>'^fDd"]֨yaQI_IT0: ndwK&tSYk"nAriqn>DyY@7`<c 9W/`5}܈C~ ;9DͰz5ч}0<2a P&#]_pRF>_lm)^PISlJ:"}|\R7Y-EC7nL :t+`ERY"C3f$ҫ(YY V_>)89.wfszM'jkV odhv?WyM/v\bA ɓ=D<;Ute~,-T!42ZJQ ZƯ}'Aْ0pWKnU ֨ 6)ۘx9&)KHH/0߀t(ĠaWj Qj=m䭳t$(a?BSy)b ( F5E%'3jԪr-фkv,C8D1SFUjƗٓS1EOZL@b5.D4cQ=5&ت}O6/ӈECSIMa4rjgԚsL!C{JV ke>jrܱK X8>|^@& Jj!yS7؂sڸC,PB.T &{q?d mp| 2_^i|='?ir6y y1|439hɋT3wg,7p]@v՛]2o?:=lf^@"FR"}RDp[oQW@dF7cw:|yG04fqtL]Iw+z}XmFբ4jOR*U1YYVK2!E b-B33I)?95aKt50F0&7+z@ƹHنWrƗ9ɸr212M^!0Ę32NAͱ$jmt1VJh^dؕF3H M+)`:4ev/9ДO + ]nm3v~-\?^7˗oQ]l2` uȶ ZӽItsr٫OK`zKR|oSGsve Da3߾ϙ?Zq>HjfsQwCDjL1Y4+ "vI_C/DZHA߽yR5]0B^PeFHcA4М܎z}Q۷Oߐ)%l9O QB<1yJnޚ-C,$^Oa}Eu~t,$Q$d 4dʰf07, 2_fqO8Ʊ7sғ | C CJJ# >T̤aeGfsFdBmȡͅJ5`#{e*n YT͞P~![WSGq/dC/0kr1|aÐ 2(d7,`KEd53c$6b!NkB35iS=qq ?KG_(fBe@#9CCkQ3 {f캌iS`fZ>c A#L0rζK"# [4Ϡ&ti# Y0Bovqac |: ~\(+24b"pGI~ 1cL *h~":9eo_e¼yeEUYÃn D!лϸ0,3p fb"W6H}^ =dS˾Ʀ\(s!EԢjqbbAC'j]j73fNAU췷ߏxSh^dXs䚫{ L?w}ܶ@%'Rx7&zK@ F!ٳ: ^.-Yp9$szuDL\B|v>R}Jfukn]}&aut+E0Uٮp<Í55h .s "'^"z7]7~dUX߯\6+b zX^rga]oIſ?)feTztw~U ifɁcfmbHXnK8P*i*VB +rp`#blfӶc~{ 9}v De$߀yKT9SA@# DrJ%}H\(V쭾@og;Q=ÛL1{RFTKNF'SHz0r=~+vo0AzrO6l38P{c0'e#//K8:l,'UJ$H|TW_^8*-*L7Rz`0m]b,%zj_J fK0Gl9dgavjk`aM---nav[ԛ.3;H91;{}9a.1R}hstvݤKM3÷9oי AE ѯy E%c _apCED4z]Yo,nq z99 98XYتT9Z#vojwẂ}Ȧ#0%(Bj7B=Sꄿ7畦t@ [vUB7jwW͡JL~B1(߸4{NԎo bxA9-/LuKӟ^gUEE, wk[zM9а QH%ƆVhx\TƒO4&\H)hNRLHms>!8>/_/^}ߐꇷخ&텤ݷ7>i;7d!ȃ/Upbk1B` 5܊BYjgG|M)A06TÌ0thďG,_f Eo/4CbSLg9%R xԃ]jj^h$ 2aNѕQQg$P.cu`!Wѵ̙4r1aAœ"7/;{k4vwos[oxqw}2K2WW- o/Ӌk~lݱƯ=n9@ ڻnebsf{,m;F0%L01_ R<@#&юq5㑗gzDǰee4bv>o_>Zc4/ 6u[kus@8Ss&v~q|^+rԨqk蘛b}Tc^8GvW쯺8_V#:ϒ&R9r`1z>thGk(%'ҁH lXI+8N3o\ ;S@^CV,c6gDc^3ءDpn{D\>eWƼm\r"}igAzFr➦wU&DW퐒WjAO/-B>Bhބ5! 9˄{akiuPZˁyQi;d 8a"3^ '*j%UWL/Bٮ2uTx@+@NA# eqS1|0yw FS't"xG㒊bN,GL%6LE|q'p:hxy Px fuKo. %V)v&m ]>֐%'2 ޖVٓ:\q""O3H<U^4Xtg#1VBZ?ߺ-2=UܬfyLANpnho=x!+y=B40Kq#e F썝~WnDEX7 # Rrx^95TsPs q-A%H5zvB6etGݤSOPP}P;[^dY%*1(;"/=J2qv[S;1gDtRzO [G&BPB9հP;C{W?.tI-[ݤ~ PNl5>A~:Ϝr޵d2~XQ}H7$3+8Nft{m nh8(Ħ:SU眢d PiטvE}bޘvI'9KxY1AƴkLMcecڽt`/4c2|Ůd+Y.lƴkL`Nuƴ8hx$8P.wQgBpm1NȰ<`1vGiɊRQsJ|KӒȓB]wzvOC{!yI$s+!7s8) +|,>]|[+֘T~g`ڊ}[v5xU$DUFЈrnR(",e0i0i"dgv[VY>'#h0mY[{Ȅ)8ҿ^rGO6G`=_so.{4HP ^Y/m%gIH J8 &Nv^VXy)u[o<K7T'"R?uD(()J|6H^7(9rPI*Y%V %FVd/=02N)R m_)54zMkOi.uyor-w$Q\8|UBKV8|Z,8Í8|!B# :T="H"HK x}E1n.8мI9 T,EU"qSXv'Fr'ˮ֎z"O2¯X$W&asY I=$֨)xʹBkޱ`0p!T`/ɭ*/Zi-F={9{~Yv-J ͧ.a?SpM"iGȵE3(QAԩSozڊp'υA᣼,Xe_qNTq/@aӓf<ӻ>-4oߗ.è1;pۯLPx`D8%8`Y V5X&ÛH% GT;Tzk^)XRP,4^~HaE0T5q&n5iG|pc6I2f yHL>N&7~{4,"\ڕEb4vB=;1f钟#LJ a *䴔 M_;iu4 ӄ:Ô`']="B` aj/O7a Ѐ 'vV,jrF^pil-](*'GOa\+?wiv1NG͠7rm&t/5& :Rj. ='ۓ Q,`lQ&AʑME|&]V[Cv 0Wt`cxaJR}ÍwT5_|"|k7VgV,P+JzF&D׋|xKJr-7VXPIi]k F 'ݸ6FнlB&a߇,ea`MsU P[ԷM{ (#⤗7G"#\;Μ.&;,O]gA^~,`[|<uiBb[{e60X qK_^ݕ?S@kd؛zt; &LdD> MBI+8J.h/+-= y!,NmXDT`ѿ(1a4~X 5xpY5MFv|a͝gMbd. sxW Aמf;XT|6w KF 5chc7w̃=49ljA扰R+‘\!΅B X@,uyǝXM*iu{g]\ Uv 7M_zn~qo/ZVq\6FaB󲵒*m"f1:} ^Ba'pͨ)Ŗ&\E_vsUy̮UcnfT #J5LgN-b5"@e.-. Uph>hcc 4-hz6b٧JKEM.@ RiVjlT.c $!@BcfUU\ڮGa) јhcns2z WG{h2OXGpJwQ @͚~z8h2Yl5 7ټӴ8~k|Dޏ$h6YkV>ߵfC~xR(7ay|8)\:a&< 1ٛ>x H$v,O-<e-+ .aLgֿ|҅/7؂nn<.Sx峧0laBn\9NG󯞃qf'Ùc:`+ "O/  WW=A+~?eBw ߧ쭋:$M H@"&f8վ筼7.ڳ1N97~I.hǍ0JAQw?۟[ b>U*Aq|2l?x>o3ZPƀAd@," l{xb#w]i$b 2.*ie;i)z5ZzAs% Je3"o5̒UD)xJ&qVJ%`h)xYO)¬T?ιW< {?~O|ȺRlM`Fَ)}"${RS581QoεY;$qMZk&"'\$͈ޤ8x 'h"ZDY?t_?7׽?Nf6n77/8{s8w~&q 8,c78Nw ..f'$7bW>Wv^:>נJd:WHn}mO O;};7Mg8'ڏ3'Op !@dU)D+Z5<dR&CD]FK:Ctu0MnNgLWXKO7[kaIwz۝6ͩ-~7Au#Uq ?h)̏uϙ١c|8Ƽ-2DF`2R5O<< eJllэR5OK U4io(]2ɂ#)t)BHU42 "ֿjg]=ZW>σo}QQ18Q8oX 俦;~/-{˷tŴ>_JDU H:ݷ[nnnÊH R+0l.zIO`5%IdJ\y}QW司CWf}FF?w]a̔#͋nAnAkPܬJRI}Ќ*LSx4:QFƌXNSenZ\siJ FmJ uH]r3>Ф)܆v4]|5#2SWmj=raRZfRXu22l({"Z3!BF +{qd6\c8TYph$ 8a&5Fu( DwI48{`;KSZVaH<ĥ:d;NxOtEE8$Di۸]cYV?za`?䪰r S筿q I}ur^w϶و*v8B^`bIgB#8E9 4as9Q"cCBcs1 2W4$iIO$٘$hfϵ $1hFb= 21":J`l"-@Nj簔)Edy0)2 wّKО0aŁSAlH%֞xO[ΡD#Eဌ$+"+pMFr 2Jb)t$%tS0\2Zj1ϜIQְ͒Df˓p %uiE +!ƧHQsfdҙ QjQa].d؏ h+4q. #.NoP>AT>B$ŭgcuI 218 XɤZ&j昍5i7LۺgIuh[RZ}L[#Ώ@uYِqDI'}V@M-*/%p^ҍgK{'$T  zN;-(bG0,،q?fs} r-ڃc})e39'$Gh(T^ Abj*5DkX54PNh"ъ@sP pEuKv ,ٶGB*AP>U >ŊgaK)7E4 Sci֟yl*mn*mnf6<FK%L˿ "SIEG 5jͰ%lҎqJ;]R<%94K5p5ے2/t\;_ ;= v/um`D{vք6I18pN) .̵erhkNdc-jEqt_ך4@Qm?7&`b- : UkUi%uA11+ *Zc7LR4DXn`,SҋʪdSAQL .4n{x[COчL}]H ˣ̹}skr %6J_f#OIw;20k^ p浩`!DŽ)ÃkALGl$XH`,a#(;SYlxJ> 5xdLג)"G-D ID@(?9+/S\_f׫4Gd.+ƸE"1#akcZ%c6AeDClȋk#ʢf Z0OTF*1/ ,P #ikڣͧo?IW7q\1# զjJkOh1%./p˜*91`<w[@TB;q#.J4NFJ^g.Nl i&8D[㶖>84\+xz#C8:Dm,3Q"x";$'|YEJ=$w!;*" JJ"eR@DUn4"Q/+b4K<ENyL)G_\ C]=^`,y8 k׸JNd7'8{EvMѬdBQ]D{pt\dZ9h;Fv5%'wkײfAOB#khiI6?ȉۢ|e}'_RDₒs9b~=k{3O6S1D4R=GK`duh1R&(ڧZLȊv-n֧bw6.\׵TaR4iQyqyA<;->ǧwwj#&o}hqnixhYOFO،0UV1 (o~dTAA}uj5 r|Ǝ'Wg0 @M\ª{\7Ts5 _-i`w#Nn %SR'Gܯmb( =bC% +$ P]KzĺX_T ))cݧ[]$>l0?~-8@?duT=-zC"?ǂ#~ϲA+9=Ys6 TZS%jj!l횶E+]+6Br: z\*JPuY$z* Ȳ=DԵ>(H 0ZMI~& Kq()#/s-H%q/Û;ƕ6RԲRWz8膧}Ѩz>F{Xr}M/@\l?-pӨFelg-?xܩ`>x+V!/R;֊PyWycΈ?Fl]Lľ!?9]dR#E.2ˉ _S~up}up{:8)v v,]TvVf } 4kr?Գ<N8lIi8 23m.u|h@\]8^E[0y797 Xv*rF8f|-}l5L0s:io5H~gGY A(pҚE|q8khDTpY7<=එ/ᲖZXσQYRꔘ&=*@h.lo'lo!߳\,螊/"9Fדu!z老U&$(}TK㨌\YRkH}ؿP"bV.޲,g1v#)XJ=U|p'~e٢~WOg+JJ ji RC4WdsM}RF3hidn&d(FRN:6 ;^U?4Ak%y9XzNy3650c)w vT'oTۧ4Z1k|S}Zyby\5_$Toؘ*M΂zK yPא,=k|k?AFE-%?E;zseʯg!!ʑxֽDrpjLW̐D.lvYtJ=DmSV@AE*U ^+Jh ]iN+ݯ W!A#OflLz_㇥dZtvO -x7d,9OTR# D$Ke== VWB))"AW)BDFq7c8 IjY4Bt6&*R!BwsU1Vְs1ZE$I@)oQ|b WDt؍"(NFVVP\C+6GxEЀ4-(2 H] Lp& h9j9jNx[lai@CP[7]+3.)h~wD䙛$&of=~06"yeǫxxhl?FdEi#_'@3̼hQzyQF Isu&]p@Lg2!nNP~fS@qּˣ>8DK6O#5,_O]Nc?e`ݤ܏&VDk]O p @+#ߚª/B rn5/ )@.>`F.KOAnN>C|;< V"wiE%Q2TlQ@}c|pRj_괖lʬIñR%K崻y4bfcgc,?cȊywYa e΍y5Ud ˤ;'pz'}j d[] ɦ~&Y+RNjh3 CdEs{Ѝ"]$Wq(&ԩivϲYV?ۊd>)~ v:E> [!y[ W0*`݌,암L!(4m-!:⸋VsP uP;΄QLԎ]²YIOIF6S̓Re=vᐳkZ9_%L i Zu?VfJ6UYrմY]5$ykbo?apmdU? mx2WD?ܠp:Oi5r9j>scǓo͡afnEsLRhdSz6G~ Ze|D&'d2'S8 ]LaQf "S'SxZN,,cl%[zݞ% A1ڱ+]od&2I>`nM,4,m@Ps:@ ;%k%0_{ZJRz$Y"LUԷI0љ1E$CaLߐ5~e⢋ĥuFI\dbI(ztKAo`2`΅B۩Uŀ R4ռF}5,S_=={繸!FJt+ymX?l`H%URpr6Za7:OߚteefUI)3҈^oApexsgøFZZJ|NZL".]I߽T+] S_k Tu/ r bo^3pXZ"\Mn<3v )6NF~\83k t5PWR%WE,GIh7V:Xuuwyq d&NFwZJCrZcW(KNŝ"O) auIS-TJj@WlLVLwm#K_ay!  IEIN߷eH: KTUU"-~+b}vXdHoS8zG%i L]@)3LCF.b4ΙD?( G}pc=RiBy+!c5qK4a]X=@[l7+i}1_LhDMG26VեJylit*aV޹< »x29 -}Z,w*t}=h9V:Nw7&nw zӽlU;JRSB[UQ⋡;+Y9qyysڲ/,?N'xIr\=|2|I n=rKA?NWm*F~k&$[zyhnj_Dޚ\)놟l#Q]'/'PmMDLID"r,VP S)W{w $sqr-U;H e;S?`zuI/pfҫ 0x%j,Ѹɭ.@5h73V, (ɹ+$IDςj02Lm=81|/5##KcHӖJnU#P9مVz~9۽{RjTͳk $!U~Oϗ @%omhM!nHb0JJv]H5-Xw+,R\7Ώ lPwK u{xpZM=ŧՋ~ͲWeK)V 49+yJLr90tB 9wٛ~[_%ȸ `8n;ғ u7YoNލe/5>QYsw}2x|6lzl7o,>nɟ{ OQhy3%]G ÌPѸ{epy޻M~{7+KןM"z~~zR!At3OM78mRHX 'l~ih04e!`2Ѹ1̞u-vҟx\dt5שͫi,l1B2W;tGzr/_ί~.=n\[].7@:lr[0ͅ7_>27@ԟ$/pS34v(߾< oN~vzůGP/y0c\w@#N9dXV\`1-Bϩkg<fz6v_c]Las9ioR|vEq9k_O^M"S?y`o TϠ| %>YAWP ϦO<8;/Iw3}}dIq+9mljLkle>>ɽ]YjFS"Y:aTY2X%3O4=Fp9[$n>,0[Lѫa4'_ҚQX;eSnt{!|NqGuqxF׏rwXާъt3f8[X"L}j3FQA0[*E`k2OLр2 =/`(F=hU K%@E㝖P*Wά/ZP|Ę?}^*m'ل7RP*C#+MÔ\MqiEg9E۸/bL c63*C_~uta҈`9M:{aI̛ONϕ4z]59 rd0)m{`TS.@J4z'T4("J$".|u@ YX,T`oC7A ʭP6X9x]\h'!὎y1$qV DdpeT?}G[5Fx*5Z=4ii\*{6iZT$vxXko|"w6Hoku'QyaS Vɽ"5u ɵ/n Mg g?6xTnQ  @Rx#ɦ@.xJX:|h}bE霁MYa=:ͱՠגHM_塺[ty.F3I w: Rxkc{]iٹ6OIxV M,t]vN>Hq׃ |e⃌`hR4bq.A 0޲oq`8E}+&(+YɈbUUcR`ەHoѕPZ2^y^1F`+Y\Ṙe0n!E2+HMf+Ƣ)E6U9paT,g &L(3T{P\u>)R:KV]+ϒT)W5D5[Җˣ.mRȸ! Uӱ 6JAŃg|Ph1ݾR> Oa#GP E>tRèp!<|MQVZ+I ȹf"r̟ aMߐK`t~#qH|],O&QY;cb-"ƚcJ(P$`{f):m"Ҩ\,鼉$&0)$Dˬ(` *Ĕ2#Lj 4h^qaRp9okrܵu bv\ߋ8omqC%-J{}ﷷnBxLfAU0+PI )qĽs|D s~ LH0IDM5IX*#Id ؞ `aus0S(eXU^ kw1T?Sf/pr*`R(w $^ fuJX #J'k+ gt&|G3߭Ws"WcH5" X%)]<Bjp&z@0(bu8Oh8+ my+9O1my0?0|@Λ8׳tPnَ.J8:`;J$ ec8?{ܸ_ji~qU?M:[ٚdE-%;=dH HVɤl܀s|_DiXRFnI …Ȱ{m;֘|iAj5Љvhn֔`챚uWKS+ŕ ?}\Œ~ 2"fփ\sU~DŽj* Kt=ٶxݔS d *0[}…CM g!9ŔzT`7s%|ѪCO) X a9''6e؛p`A1q] x8¥Pq%ieD8S RJ.  phJ'E1:_ҷ{C$ۃ(#``A^:WZ 9wQx^-p4c673r y^wN%~!nct14.6)[oG3.V @#Y"hBq,N[Am y=– W `X+(`AA (TJ3$OhúL#M(Bg!JҊ{2;lG3ڽݎI`-<$CP=;[ ,g+0AWG!w ?r̴C̽'t |&6fd+St+<)ƛpet@UPMJ{VUp#200źެ;Q\L> 1MY\xFղ7|fvY]!ɖ]/_g  {C!Xgq-X@/&?f߮bf:[Ew1Y?Gx2[ 9oF{w{ą0n:ֿ=P[NҬٓ9[}.l5>[[z=zbՋ[ 泵ޝ:L3Am =%YjgV/6pcT$Q vd;tQܣq4GA2?#y٣Xp Dcj)\)}n{WJ 0oTh?kFHJ5knQYҀ}ַ{롖&z?aWj n}sQ%_`7G *dͶ^] x͋zO9]?#dF暗&ɔ:YΝjTm2P7o[>ޭS-x9 )y|ja B§O-lRR*XÞ35%sE An?j"/!"_߻a]I_>gNr9Òs\Z"/~rTzI!=9~B0_~f&E+F1}zl8 FI)!I%Ϯĩ#/SݼҌ zkteEҎQ(rg̠R&3T"1HV`B3 ]aU^O}1|v| '{ްԀ Uǟy8: vԿa_q:jF0w_,yDIj#"dw#>h2TA#R:)a:ׄMb}E ,xs搔kn=S|z[v"x+ݸsvBԜw\SsG wZz3a{;/N_w t)•" 2u4դ 1,8Y("SZ,}Ms&ףOK uSc:b{|S5:K v~Azx t('Ix̪]9fU§h"oNj \S@;ѩ> }gA=Nø`_`܋btQ1 m~0,N6uԺ7;pk~E*x 5v c+CLZS lpJz&۪^YSqZ#_>m^Mkix7 뛆taesf°2$FJAZ1s8C(eBB2'W& vWWꇮ&, ҀvMmǏoMxhƯ X7B%2W̖̳#O{R#*)Z x4G\Hr lb}_ @|Q:D!RP"DIlE%2miug7j`E- N9& ]Jy)V00; d %b M50Ec%Euz|ŒJ38x2^M-dWQ]_{ G/:DnhbX͊{ O)B(aOjUNjo0/qHt0,ோCt'x`L P)uw`'UF(q8haKpBІRp(Q˒(ĸÞ&vLnFHՐ5CF.[ewhcQ VxJP&I Q7𗢤>-I685ZPt(=tOhq#-NuHWo[PIы"59LDa+ePNcsnJPƩ&c)HsY ۽{5|)S#>j̝0W> a ɍ!=~2.ȥ-| R4ّ:dΥͫ6s0ǃ|v.Z|k' :ªFvLszEep2+呸zƭ93îpǒQo2+Ù {}6€:_|4͗Y^>:)g߸U5]{~"!AФe)SHHkRkmE%:SC%+JH.m,/i_w'z>M˩Ϥ췯LH%Eٚucq, g5Ὲ.Ix^ʄh 4IKBOD8z 3iЇt[[P$cM~qu ~j:{mΕ-,X~--Bp9EA#  f6H2%Bk/r$*%T9ZNE_s|k$Zw TO] ,};b4AL#X:7A=Fyw4~ѶsaL_`A,5jflx$X00 _dsTӠ1uhPiZk wti߱0/%L j3jވU3v^ *|ORSJ( wO SbL1,z"( "B9ɲ!<9;{)Dj'!ooN6{P@zx:P}/p0fwx&%BO4甋{5{$ЧsUL;圓CH*|- ŕvvJ6ʸUƵ2}q]2Z\UR1TRU]jm$HlKiǖ颤EHANPvǏTH,}"uX7 9MޱnNnmM_in䓛=t~l3@MNuZ2cXL:e;jtZmpq*츫K0R\hʭ1!.ϑD fi s( :p2H[+|;AaV' ;eYԔ&" X )O/A|Ft:;ɅСK*##$h *aߑ*⚈ENZ \fHAJ)DC\8奓"`B RB[X;HDJRu?oن}O6NCp BΫE#HkE@뻭0 ݡ0JfB D;0ƈ &XDVֳM1 @T=0ظPN!3{ӁMȟ߄=džgf`_էcAnƨM#c~e1]{MKTɤ'Nc cĆ4p=]J5>W- ^:I0,FG rBl$vÏH*:aa>tqǖi:T 8g7#xZ6 >㉝saB%}mq4\~cݝ\Sj"fL*SY)+TȠA%VH_$%TbRŠҒB3ĘaycFy;EJʲ0M5q*J$mF0eoJ +sGFᜑ2͆і"޹p[b 4FpB e!2rZ =XX)N.ҙcQY"O{UD# lta-+(TI%/ F1FWnkEhB ]F=ʱa*K`(O5*9fena9BF< RK7K`Ie)01.88iw)U0abԣSe ƨKZ`C&L:l6ь1&Pš J^<*"{uͦKXf%׶Ξ>چ˳.>STb613Ji|W pqP"r;$ N 2\ޛ! d h Wr♝TYf-x!eS29?އ c>9 7ďѴR ~e >[ZS|o 3~]dnӃr #$&nW7B w~>ht#Z5# B B/aF)B˻q&5O|sm9~0u(L3|9Z%kdR}-;]oِn.;b;voaJ)&@LYЍ ݣL ([3BP R"aF`D`훐ya15*DD:j8;*rT )W>Ih d/7pƈD T:M961gl ZJ iUPM%T+aUJRadG3p(|3buQ^rrC- FQ "@t I;Z㍌EJ~}ؕaX ^o`*M夢>5+$wŃY3zV'~Jz~mwv]{즗dj֕~5ʕe#̩ާ^?{6r/8sfgpyفA67l2栗mJ&yI#".]U]UXߺ/NOޯ< ^taR{gP+rYIC^bw|M= }F6\0dZi ڵu &׺АWA:UX5ܲnK6 }F6*BTpnZUN 4J].w!J;':]id6X\WNSP:;]=eʈOW!sp5vj]d6~ ,ŲM^4\Hd&l(RI9NtFɕmq(-243Ksr M #IRHA]Mp! &3k̓lWT~!ƀMA k1&3,;y1lP ,PH. (4&J7aRiLj`eڹ׈v5QȇqCYCBCu-!.B׷1Qxr|o!{kĸ^qN4'B1+[[*սK~D%*PtGMymî^/kyQj-b=@z~-@sZ"6~qKE>]J[{\\r>{ 1mY43bA,r>qł1KDԈ̖ AJsjDhfJT: 0ODIuJDԈ7X6J{ERMebTZh]|T5*.a;%=˺0IT IR7CcAt1HE$0kU*$zDUS3eLP֪6\L4ۆh |fY/UXS~\8Q_fO[3Ue.əˈ$-yS\%~.j3]q}czMA|6 N<2Dhr?}sY @%F,:M [0G!-i%x;&h)4N`h«t)1㵛;69sf}yr!7@$ 0ly(N 'HpAHi|ma=ޗh֣ Z#7{b'7%ǖHchoil$A_ch9\Q&!Xժ揝ux5دG3횱~IQ'Ž^POŇYdvBQ:vD} 9R9x>E\JWiA2@uRX0rI`ja&Ls @Y$(ƶ>P eglʱʦ&B:ϗE+;34) ps 2_OnҹK/z4g`ii 7TX+RoXH^d9R7%cgwWD Y}dgq~O+Gdp?DC6#"xah$^b[  w}iVI~+RB*X㧫ȻZ1 Qb DF,I2Lh"aQ 4[' 4bM,ܥ+DI1BHFi ŹV.M$LZȢ(y2"|uA"4m\\&{tzgSMưIHSkUEk(xMVSa)"R;$!E8# I"U*,2KF5iԥX7yF2ADo?dՇ͠!Cku!`$#djI?T-ak$gP4UOn_&[0>͘*%bUl!>~|]塀˻×I & wfa}חy=~p,'ܣrCYcw&E~YV"e/ 7cT\Ԙ׋5;J֎(DF6FC#6ȳ{etww!1646HHmW;vQSjd;;Gany0/}# s3*hLDW>/1DIs40S[>3F0_tpL=br/*,(\*͏(sݠc?F$!z]QS(U6x#kac(n 5>6KlH?:EsqotIKS"W.a+W=@ NaؤS̜Aut9NYϛ^uNAFm}wKGe‚:)B/ j5<)Fw9pȖ8ҽ-:‘n"?աⴶ 5Th`vy_:ZPOYZ%TprΜ h6~3JDGp{*1:|\^PzD2n !4eQD ܸkI4"DCʂ[<7Nm %m@*f(amsY\bDiDtk3ËfF0b0;-RQE!r!FSpM$&CB±;wDRk cdY!ZB݀Lee3V5p_ a`)p3emx1 fAL{d*VyJ& +R! b fkEF.slM0c!Ȑ D7wЄ4+f[) k )$*n0.@9FBJ,PN2cY!bh4eZv2^9o3X(C/,)\ f$'`rXHuX#LJ`!STP<\_ݖ5NQQ#Fb U# ]T0QRAm'މKfh< @(Iԃd'XHB$I*mBd#QlFDzP1ʉrV#1ԏhHId :cXlZhyj}͈:>AefX~?]ke _>E cz̨2YxPurZhKL:5@0^f5Z/!TZ ~r=I>3WGۜ%Jt P2j|רH lRhyRVK*(!$k1ea׋Y,D"7㔋4c4!%܂D0%!ٝk<ė\)ĦxO^ldw.6~#yV%x}H8+CMNE5/k$1C? C[`BR$X,!u}/FoLURU>ºVtމ>M[(yr&,Ko4II$ɡV1Ǐ=RrK6TsDsEآ(BncbVZ#YUTŝTXSN 2J[iIkx+}&-[&?ol%&߽q1qn5쮎*F*M?چX౗Gc[HW69@[ 3qa"|kų S˾!GwJ `G82_6hc 6Y҅G;ǁݝIiJ)G砈ٍA/.y.yYMln`׳D: sS.Ҍ4N&3뀄3.-eheE.2:QlaޕvkMV#l;M?P&u Χ@>njzB S ?ܣKjkjw(BTC$Pԭ9O7֭'.ܓ"%f_3ԩ91G#S +sXxw*s]R2\xuM>iJ=ǜ<2c,{yZ 3<Oi:Q`s/-Wz/6ߒ+/|2)sOef.AZּ~ N i@0 F+>39*_i&f~m¹&ڀ/Ѷ>JKq;-3=d3T1@ͮ Nt,k4U,B%W[Y3cv Nhn~lj5OGcuKݤhx~Í sZC_zkK?<ÀO^5j6O:3K'g|Ɓ}q6Y: ѹk_dzj~>;Wxf:m Ͱu#]v{aOۡ_BkKqhI\;ZUg;t(KnR1ڝjnnt./rjh16{QQ߽ބF ę:E:/=Oq!Eҍ|./~I2pqv Bt\O^ >Br)RuH 54Hܰ>Q`s NJu^:PH3*έԭ$*.M creS &*|M3ғb*w)򦆪\jjpd*lmҌ,X̮g^]~+j5su \'ƫoh*{O$)9S@mav3Dfӟ~0]0Yfʯ ->!/NjllGgX4,NRQ_Q-}B(/e>J:͛~ Af4X0K%h<)*ȅ$v& 0[>(1(oXP3e #P9*IcEa5T EhsB#e$7 R1TB0*&8Uoz6S.z=u4T+k .q(쌕8,{I Oa4_F2n^"6 nq;ZV{o:̜YpL젠RIrK|dF>R+8~7R+[R{l9V~ARYsDL@k/Vfa^F&Bb6j'gVnq̭VkeR5IU= rV'dH 0aXHfA$@*(C,C( )9$4Rf Z ='/z; e. : ,yVѾ~jxY!̩b:\NpDJc Drc&<+mtd("p&5Bq@.L%LӅIҫfۻ_.w(ha~/?qJI8ى{sYm!C~8Y&|3vzXϽbZ̓^R_x\m&]m&Sqf=~uS(#ţ|,Ѵip8_goM [2>dJʧ.0Lw {W&z7y^l @[L Tv$DwcA횊Kyj"{ 4 3o'@_>zb6f= 0z)al0~>&a[5wC`@Ą7,A]!.W<p&Q% $aT@4#EɕT8G;3 N産֯L fچF',à% vȪX~ECDrVj+(cǥIS"(/9cI"r.Г%>(vm*&j~'B?@$~ǡ ) A DcyF>B흁^Ҳ0 {u_NVխD|Y;oJQEBW@]yJgX9:@/uzmPcڠy> #k?T9gxIU-ਝ ^S<$&%O.%BykC"*U ʢBWWf-t4It6 (#&k*a9KiUMr M^[UfֺWSe@!9EF/SƝ~\/ŵN,W2.U+/6x U i*zi~9Gs#C0z2 Fa.GmK[,]t?iOƳA7D*x^OI 5H+wE우b6Rjsn B X>ZRj< !Lfw6l~j@E{a>$rʶͅL3mkmJvT˖Dۖr< `0pc}>X}`!+/C ) q!DH28ҌiMPHbB@T!|޿-5 o_o;] Җ3|ϓ͡Y%cz^y9(\w##LǪL <`@#,RQ PT(! R4HaIEQgB _C]Ӎ2]m ~["dl-e72:-fS!3d^^a~{X-- Ve bS]K1Hfp}~`}b|ɍahb;OOoƪm0I3o7;c\s2(; ?SYtGŐI 5]ܲ;!Sĉ|DPJ`ƬK 7|&ÀIAi?YX$QF"H\#蓈(_D,uEp㽔Z JDZb:0_(a,>1@ 2bu|Ěu7fQL;P @d)MbJ2.>#Ͻ-zʝQcT@}ɗ3j($@IعaYĐ`8P2ĸ愫uX]LY ȃu"F(u(pX!N+ȷU3ε+,*Y(zq+rDjR."VW-z%"&1ՉXEtOr1\9\@:wty*97BeI*M_<i:^47Cn9S e1O4 x|fN\X^"$OtRRV"Ϫyd%aj]*p<2Fx|w]WFaA.=m (Rboevhyc~G1:r,d-vFK6'ZξNOGT0lxg#:LQx&<]K9޿Áh _Fg5^'@ TL}bUS T> vVcnuƱ;A+JƺpEb֚UDqܵUlpk*")*ȝ~q;=rjD%}+U?C7KtW jN˟iV i@Ґ >;/́צ rUڋ́EӶ-#Ϧt%-B3Э=ޫՓonL@:3Zpe>p4[+K0rfI2bZGZ@+`B1h hB̩`um*V>ָ(H>I>bH 1%&  !$S!F"ҡcjM튑:_f*i''/z HQXW͕TP@(U@&*!QD *H@6#0Hd8Bn)gm* ̦B:_~!ٹ"p)O$":UpVF0AvZRnh6X+[`=fSok6zhQ4վ\}Q( }eٓ?ڌ=p ct$>Kw91)hOȞ$ۻ#9 L!v5_ce" aA(6_?L$M6"dd{mr8M*N#*4R'x$q;P1[|d8`JY!A L3zJfD1eE Oq߉E0 =~DS|DI^lKmߍ=%[Y׶Jz^/pxjUs/^&, Slf͇bg,X`BܻǻF2'g%-Əd$t*Z;.SAR]}c I^˔^$ Ju/${WQ0{7N&NCu*W/,Y{ @. XJ$ZUy:Qg\ x,Pc'.ܩz2^o79JI3#ScP}BGk)HHpwd(dw?JqsUc6 @(', @W$ όd-gqO` iB<]jo#1m= 뵫@+ ò4+.Īxq> +^8ծ\ڡv4d{>'<_t=Ӗ3*[ο6JLI V]-wKVׯWMMp U]ԧa \oפ]]*]r6dcZL6f1BaXqC8↑02D [t bX}S %ͮg!K>V{} r?uhP `p!] wޙ@8[ tI bj˵YZ수byC]y1Br ڰ( Ni$XG8TfԀD7-cOw>@`jcrT9X/ r1Az̴ !4, Ph(hD59( J#H77^5*ĈP\dJȠ&(c10 B@| X@PHvqːG! a@V @!B"WiX+@UEMRH$A"p$&InW`f4V e3OлTlJmue2_(ǹwAe76Ӟ̯f|l1 Oo^gw?n#_!͎4~q{1.kw>dF`4%;=դFaj13 CUztUwuhG7*,ݹy=!c %u>}Kg IM_&P}MJ"L#P6+UJZh AP;4թ%N>517C!*X\;Wx4']~ِrBߧ^'o2MAo cbA^qOx`CETXg!4wH)ʭdMhAwܿ^6L WNam'IAXP&[ ^=<mcAk{Gv<M- {A43F!c 8\PjXcXf$2.T<w>$s1d%) (b I˼FKʼڃ}?vs[C W9b-5 ,CR$VNjo3 U|Tk ep rka"h vd4sLd s5X`rѦ(Tìe4vDi k2lXs'a*=|p7^HNI8am]SRA+qg8Ƨl֏VP=&A{Cl:ؘ b%  in6KLpӓ&KgVȣ*kg%O*'&<*iUlh|>ƖȈ<q8.1# JauĖ݅sɮ:|ڍ^KePeSRDc]>l}FZy+wO9H0~Ħc~bm&< ;D2d3DY;Mw[$ar"]CBf(ǓG3CZ>JH= iޢ*|2Vt 4`GllfnvPyC$s޵>{6"9>woLjy!0wX}/sx>TK8lJe\c1hr4'y> qTh]]c?A\"Aed3mh]&`ql"BDUh4Y0ju%i;ҿf~qu"&ܬ֕CHa*w6:fJ.cg1ykx)^o#:ow])LMg2Xe60-hKyaѶl9(ta^=(13Gd?@KKr'o+oPnhZ\,/Cnͨ:#B4f8ȺT"'',4'Zz[H:¾Vɂ20Ȯ<- p=w4 L;Eu!C1%csc4v4!)!HX.}Kz+u\"#`l?q jֻJ0Lk xciރղuc Eψw~:נpO&ANoulMރht G_0F?~{Q%+/YfؙU=Lt 3v6a4WDke2<(> 1bƈLZ=BGiٲ!j…*vSlJVO,bME.Zq)y(]t1d b'4e;j1~4<ֺR5]{joWϓuw$)'.aq`W;oMS=Jvd }C@kllךw{R V礳nsmڼ ®WrI?:p<'l'z[dKO gpg%oSsiպyĪ/?9YUսoؽf[zQJ{Ӌ.* G⍧$g"ZptxID%:.OC1=GmǣhW?_si+?x6zPLhA pL0P rB IP^iߌ߫$!mA!MԔ!MEΊ7T0Fgؐp,0b)F-`4;1K !a pV<<*Mq.yR(~Ond_ӄaɸ8Is[(ϫx&W'Q-yV5BzK=13r)m0l!u IJK!iGA/tmAڟj-u [c*qO3OnO5ѢS4zB13.:X]fV)-0 *KȄ@bdrj^90sQRME=d|әۉ^d`!:.|sMnm0D|y"u>φLmpuAf+ cY\H}2)LJϋKXӳv@mp(fnS?MF0ӗIgWL2_K ݥ(p;U4I={ݻfzT N;xQ#TY_8wKfwkCC&TsI^QM++150&Qv<3ʔϻ%3Z;vkAC&gg?wR1:hN82h-Ѻޭ y*SwLʳL1'Qfzyky3TFkg-hW$7v;ZȚwsnQw4n'\Jz4h]ֆ2G?MS?9 ߂&|w8)ޜ5ez~%[9~SVe8>s *& pſpu~E_G1w^ ]If~G+뱻6z[Yv[rׯ'9~oYt]o-|Yy׫m×^φ~ynr<H33 VQеYJt5G.:P,)9 IE$] ,(͸JB}d8S^PνpP ۉϣnHzS/4{4}Y#_I> Q@O]D-I"y='0ƚS)˴q[_xrcW/o?| סzW\ߎayb^źMPJHo,=~rv s\~+ FŮU ;l&HiZHH zHJ6 ňچHHufn@)/]u3c?ěճ}XO3i%jīo~pXFݍ+D :@gfA&!Aol&GE<MWd)d2 ߔ恸L ҁf"r F w^"ĭ*mkz\p2EϘ kXS]Ef4䤘[`UB!f\rcrN̤L ֚FkX\`e}F*1E-z]²E {b5&2^]z`YA E3`[f4Rc֍rF"9 E1 m9F~Q $d![n2.BWc+aBbA[;WSqaEjL}FL\e ŔZZ[e6qLD+|}W, <1`Ky;懝Ofjϓ/b\mDby .ʴ/~f ?t]GϿl'qךL3ȣ3m1qgX&s[A:d֝GN++M*H*Eμz!RO*{Z5p/_ь|Be;Jբ r+rA1Yf ʌAyJW4`b d$l˭k>G[ډٮ(~M0vks{ṽ&PTpAU͚)!%j4QYV$vm%v5khvZ;#Q!e-"hǎW*xw ' k3H$3nd 6 AD4Ͱ"_rYcfMպ9$;!cȫ3G)3+0VF u--Y"|o9#ໝ3ݳN0glr+=x.zL^5=[lޛdkFgczKE p\ads&aGH3$UX^hDnG0[p*[3}g0 x6`Z0VjK{kƎA~M|?KƊG0EZ*q5'ϵuL'ʋ y:m~pPnGp{=:BXݕ.sv )Ctu*-;%֬|'' x˧˴fEc%)ŇXIc;9SڃQEUj&> *'@hP G܃\= dnh㢹HǼEZךE#g޳,h*9&*tf5{:E5uQ?S_q`BHBӬ5c ,F>68u#o&onnÏ `6wC'f5h00c? ~av;DpbGGFc,ARּ ^@ kuWneѪw&NF uTNV}~؝@ T>݇0 pXCC4KTnJm&#% rjixiQN?,ips .-9ja8ka ~林>ζ>ب;~tXΕPiR=e_$Qj,0+fv"(w+^7̺2pcQ~m囹՘b|0l_ŵ? u+cX`U/R&Pv;wHuTqOQO3`JPx &6Cwz}=pRĦ 3#@$R$Uݣu+),AaqV d1s+WAUZDptsDv>(Z_u {l^rN?rsbt^(| K܇0Y޼ UHBG5a:)ll}?]cCُсTyڍ tS56 "|TΖ'TQ59iVKgZC ʐRzЪץ5I%x ~&>X\ ذ۰i8$Zae RF(4nOĔ3:EtRNv2 -:f ) '$*mQUVsq$X%j8Tt[o/&+HV'6=YPBi\f"P3SQj)#yq}>xQzq:cNcz7xT_}9?fW 3C]]PA&W'p6Rqͳb*Wox|pRG>n77w30_|4\ v+xO2~Mvu:~A>d^39pȺHRˤ6cLf+lenp9pB\1JsV(wq YU樰F@tx\OM5K"Ġp\yy.,ӆ̣bf9.(GY]v]E9vY:(R.I%) 0 b!~ƆW'^#2 \ɐ4V[Kt-2-f~2ɘ)qVP^d.g:17w\ x,*;ZYH,ENkiGLӴP_c"וA=ϥ Ҙf"c%fZ[FTGˤ?%uH9\(9fL9Yڜse ^@[9l@'p4dٲ ;UL$n=2ÿv&Լsr\3[iln`{dXʤm<,@IDօ^VX̒nиwkwʙnc}#@g?J l=zduͅUP?c.hg)Wf9bY=gX@嘀ɚoG藟bZìYgO7|@ffُwңl=;˚,lu3V1A\a?,鎍ǽ3'/9p"O[%#h@痟:Նn)8"Mȧ렊w@fO mnGbЁ^ D} uO=ګOܬ+Xtܜ>`Z߇5JFWyi8unXԎ~y4)Go^_(6<[?Mr֊͖fj|rF}Wҍg˗2J1!~B$BG_iczz#{y^kV5]-k>:\Ļ`Ĵ!_ih:ܜWFd߳Ԃr" 16<<2Z -jԉ9-eHƭ̙2r{<1MuM'˕ܩoe0JDCk`VdkMW#N4}"ѝ䈙$5I k@Tx|V9;;o~VΧw>ꯋ?oCrLp{;{;ÝgEً};6hߌXNvJI|S( O1܂E~lA~OV1k5[~Ͽwս:HP6䋟]9-`:Ԟs֙Z0[9\5l7/FgnMX~doG_B'VF\xL\UQI`PԙY=7=$`x"d}UG;K{Pz,m[JAݐ;6B3B`\d, Z)7e\sggG56<3L6$nFAcټ֮\ە ?l5\by>sr`,ɼ_jΑ͜Si5/J$^2;2|-0:=cjE&_풢%]\!6\cCcpcC=zp)& O`@v h=zpzH)[G&$mB%h_[AޏNQk>+`uo7}`%[f_Q,lܸŲ0D\0ĕTD KsdaG.&;QmJa tTbƌg &c[VaȽ-GwwVqZ+kࡰ6/"MQ TRUX+ bDKKB`'8AcBsBP/F%,I 3*Q.҂v^,mWC$53YڢA g5"V SJ?B=EPqOz1*Djv]ÎF)flF ApNJ +0/ (Vc9g9KF!2 bdW!dOT($si(VoCֈb(H!eף [(a!(/"Y ^QRJ. c(vE1P4O46a`iCޣ,Z->w< V+\9D˱:nDFAQ" CL^pCR͔qKǎV~ Wrx.5!a)j 6\w@~B^zW-JIM:äwR8-PpVb9'L`m 4p`kI/($M<K™4aDp 8dqKR-1?OOU@tˇ7V34]}-(Kp;w5N5d=lz6~GndDWwmҫ[ rwNĺM[ ԪӶkj@{Ґg1:G֍+y-!6Z fݢ'jАgQ:@?yߺF|B˃Չ}GvʬG.hj=PևٳnuAľu;]e=S[艆Z>4䙫hN)|Ͼua'n<QwnKiLںEO4Ժ!\EctJᎱ[4RAJ4ŝnށ-4Ue|12.d˕iTҳ2XwlZq \cK MMa@( '`L1 8RץL(F62aJ uPNق0DIE後aL^_Q-'<1hOh@|<%R&Г!\EtJ${u"ytJ'.: 1A؉$ACFh"7kĹg݄:c<Qwn"$FY艆Z>4䙫hNtݩ7Op* p9ŸZ ׺k\yekI;EЩ ܼN5V)B,V;8S*9&rx4@ڇz֭ y*S볏8gQAľu;eA.vhuCCv)]J?U$ꓛe w ->/ƻlOˇ@ q$vl~_*vBӉa[wz)xDPI=io:xx=G I{#3NNn`D7mЍxD I`78#ބ~-naL,H!bE3y D/'Yv3yɁ ⽲'LJ%ۆt泙:Lf;ۮyh Mw#҇i0zV"?s߂b$cfKAn Y .i&02vyw9bkMW F?E$n ; \Ov?{Oܶ_ae܇fWSR*\hQ o1$A q9"=}MMW>pB/3ȏh@̤9ݞsܥzxxu'O c%mA#}7׮r1gݪfu@w>ԟg`zX:Oy$(|='ȫ7u`y!|c +Ȩ|~  %/Ξvn )O3OI.*AOD8'P CvΟT 9mDc%DCPފ׳}ߣ)j㫣EHGF[gf^<˵TԦ<1IՄ>7+ yKiV3e.a!M 9 JBSLtPxیBoE}e$>@;(BJ:̬/#@HG,=NyQKV#@;W4n`u`E"Xp1U1 J7)UfHR7GuJVs5?+ցHms;O:6) :ښǡ55ʨgB0MD9)\qbz6Z`(cp6A$K"y%b܏a2]ǎ :;1%3(2K>=t@ 4Vx(Dža҉-BM",r? {qZ{1BQA0\bO@2ZbW(ٝ8 3'~ݘcR8jRZM#E n, onƥ}0ѐNr`(GgLD@&Xӹ, ٭ q@j@F} *V 7Z8Igo`O8SNSle>+(i@@{N) &$}dX Fݨ"븁WhL̇0)I{Щ+ td*tq}}k13CƱ ;[ que(*$XcxhT))d0zU_>ED3iML`W^k~4S]&Cr>ܺwʢT0ב̵8v|ˬ>PPqm l E.' "kZ 0a !<>]o45>G.vg%W @)tu}`]g5@DAOY { ce<ڣC%჎ׂ֮] R$_y_/XPF#r2M>z!+c$Ҧd:+:~V t^D˅8OVi{ZеmHQ@`l uHw<*0϶zNИBF >qDo4=鯇{O۟b]!% rkg D""P(0PhQ<,~W{M[#xpsӤQ˹ TRk>k5oFãp] e vzs+ՕiZ#G)Wd]H<}O7=آw+Qgz/ $1l-gABx~01Fj"'LxچF mGހ?$VPxM*MB AGbBI Iqp(Zَ$=PAR.1Z?@Rkb|:.՟DYUkEQ=ٍ2+e.Gq$Yj5qYqR!IJricKŎتD̻ ]߽sٴ$e6m/ȋ/iyY#%="%c\e#l,3rIo Yv)X)L(6QƬ<}6)ʖ(/ 36mJ=<^3O8ҥ݄ݷͅmc-bs(+湤 I1Wk~'2?Sߠˆ~2rdhK\`*$2'{*Ό'T!W{:O>R'w;N7ܷrj0}\JlVԗmH*qB_@I8 a̼hֶ N7_ s%%LGќKm2HpLub燡.ɣ~\sGx29tFFuO#.&˾͟Ivm)%a`.y/,cf@ (݄8a&}xwCIos@GVZ_sLQ*jeg$䈢3jJ5/eR:g(1uDW=v]H3!&I.M5VR5hlb-\WZC+DM>^ai!ګQR&@g^#suЇ646I{ Lڮk Se0&}5jkGNw2$rBMp =HQ㣵3 VLAN3,6o7OI$^б:gwl˖YtYyDetJf#ietpxҸdì گ=9)>v_`h/(&c4gµ)H+wO7ëZϋXfv򾌑onWDRX bbHx2HD~N(|>A9 ,eEHʊ1@F'LƜ3BB@Dce @c8$ɏSgLBmEAZCRG ,9MzxAc n ?o H 4Q:"r6A?BW*˶^?-x0_e~=Q ?y!!G`Ęy?1DS6`+ j<"r)V+@o>y|i ]2#]FNV3-tނԹZVi oqpEaoiO =_p}NEHPj2D9@ĎlNMŃL[ T>JnIDE\fu ZQY;P%uRߢ7 f|eP; m ԟ i+%nYtcoWpU5gg?4;LtkQ `l %!XP z+aI=plA{1nmc;8MMysbW|GDI@.cī1;>tD;9s49nxɛF uEwU .;9wq;f:_e-)D\/geuAJn/jQKDYnxw4Q7=*_?a0 (C:u- C)'BiJ,V/G×1F#z֧' (/dΔ+Ä[,\lSyX Lhqcd~8CIeS Oϩ c0WT)|1Wdx+KJq)'K{ RgFքY2\2/e|͢c[zHѫP,IveXf{C]:cD~_c# 6xv}%"ΊתHA=)!+}Θ7M0 pAy{+6:>9*n^υw#@v]'a(|ꕥ15{we4p&2Xiԗ՘2??X]T_g|ݞ61vźL[j>=VrodE5>Cx?Qr5ϖ՞f9z3e~(1=wn)TǛ)jnSvs]E>%.jJ9Yﱟ̿$Z_}qXb+DVYfon4)Iir4x7[M,'8ϖn"۫_$s-rޘƋ>2t_<&;%Mi筗!C1U2uKe %d־twF&-.lg.}U2WeۋdM#~=7RSR Lv^)\tF޹3tD}g3 $d>*nMbMtڞ&g\L.rWA"=b%u|mR}nN*dLf~}ޥ?t Oel|HsΤh^b]NcyabSٮ/pé2vBEQ9ô D""hs%(hBqDp?M7J\C;,ܣotd,TP͠7sv jNR~{57k"f|2=UfTGʓM8M"?SՔ22L),Tr)MkIle4O7%[kW\rЕ~x4WzB_R9_'r׷6jTEr$JfT‚8gݘ'L|}ӧ3SEwْB!,ښE|pkْa v7>Uv.~ME5q@#оLTwo،h$pssqk^tC(02춒l4i+q(58wP!9 31zBl bڮ^d ˍDCb( =C AP"' w3ipy[eK EINe\L(EgG LMʭ"ރqa1d=H{6vVN%۫PI&IUZ1cv4n!{䑘&ֱCD:mj4"C bD@_dD^`t9hÕ{'IA4DҏjDJSAL07>B~q :-@mItHZ z8RfjΓuYXh/R6( Lx Xȉ34gM &8(H (kk-AO~=Z=uXxCn,1mE*_#.hb]z*Dzw_*zꦔ wNU7InBcrƭeQ޽w})^fK:l䫾w2VqUJQE #{E-g#1у zjaZhfF6It0M/-ۆSg:!.k'GX`nm: t;sǃwZrAv̋H+Dzg101ۢ,j4J Y :IG1fBsv5/(i"wfVudhhe۫SEz$+X ܝOsGѡ SeѡQdH-Dp.}4!z2HH+V{3~īA)XႠ QΘTXJ{rn!)HXcLݻ?&}ZIv|ώFϛm8ПIXS*X;ه]PTMvWRe~ZǤݩ̀%t<%f:]p͘P}8۰ї3Q8k Ny?uـ OoiFrP4/ڇ.Ŗ?6, %uKIɃ7>:?zjA!,qԢid "%gxФ0_~, I{*,-yzgJj땏, 7CBٛ>K̛1 V3wCf95M7a2P,(B*Vog]p7a_+{4=M,Ǘg5OD~T)\ y>F`'EzZ#mjOE*!vRvL-i'7v{Kނ'ݐHtFN5rwi]9oO%tr s^iA@UKڒMI[~s9ᒜA5/#uJJ68Js2cYd925䌏ɠ^Rp6a?vqC^e>`{Y=`,OGEN D E2\FHH':Qh%s*.zE{Sktz]ޘVVzSt t+ bY:R`ZWdQFsԚXKڜXye/.EUy$0A 8P]r\o< PH@ (ykuۊZuפfc C#PWױ%\qC4PSjMR>Rd--XZB2"=Xr!zfx.; aSe@҆Gޤ\1};Ѣnkjnܢ dYpɨ1xg;HhG)HTڢMe!\q Dza ]s֢hKZM ~4 ]-rvlysLg֘ЧkTa'WZSU` d*u0~su3N@ {mOIu@g5mq/Y4/h^fѼf5"xϝB #=ђzfAXh8Asp*99UHl!t5%mF|ӽ W^#!b3˨E5M`G.هK&QK 8t7F䧶0vY_FS">mX?)Ƙy鬗kZ:DOP"۸$)' Td90ʈG'Fs*8Oq"h%}nuFctTS9 J:f(R$g ${d qP!M3]"D^9P$3iU b75y29dZ E>OSqh*VlDn>@;=gJރm; SZi% *)՜oGtMY9K RJ^ AfU9v[TK#zDeNnZ'c^-H%c8ml]f2xUa)T/S EXkiT9X\g( GnTh[tpP/#>Q*3\22S$$8W&K "asϴ *.[f0d8$24[l .5ޠ=#P"kgd"gD@Nh2Cm`5,u@ZTb?K8v$hLv Ɯ줗Jv }tо%*8u2is60"X`*ܦPK$sfnT-M2Rm.Buֵ *;42m*2ҵ`Qd^'Pgeq6Ӄ ʔzPԔkE;TS55azu( TĬ< Jq^/NY5pPQ5}H6lɰ}{MtMCv))R"%[}OOtKkQp Poj6?߮Mt%$gsMvGvlɡRէw@Q)݁=j6=A5`#nhHJJ"<݃4 rT3Q Ǟb x)<4CpC'pnjBR>J*dvO.6fMq\ƽ=6XUY'lj0T{}^Cfh|;?|W K_ bTvzFv<$뚏 β4bU\]OoZאhCٽTF -$/X`Pl1j2~Ym,P!ia8-rZ _waIo.lqv;F&\5.jZn4ߙ^lBuEKؕ<Еo~5SփD PP,d>baHQe令HXͨe8yQjpBz/ a +c60eJ~ctܗE'K;")\4DR)',јe )PbʞK7ĩLC]+ !0jCVko 5RJ˳zhJjеА u^)~2FVW6|PGPX]8?0Qju2 )@+p6rAaMhbR%'E;.lWS8(2瓋Y.6r܀:-դj-+&"O7(4pb6ohfq|ts.YE%l@%lj_¾~6*܍s<|=u7aq;_C +Ph!5X&%rfU9>]o`\|<:NբO`izx>ᗅ{%pWՂO\`벪,f_wS0Y&/ B{B]QxA.=$XCF òԧa-uHdW2f"9RW@^97/Z v׻to_e0p?gJVU.`%- m7?3E2UW*k0W~\5zx bӁpMy၆h0 I  ilsf|4̴W[.g7]HdtЯZ/./ގ;M? E,?:*dd"~6U1V,^IhVGO`JpXEm2j}EqO`MIЊw3tǑcC΅ nvJ[&Y_U Qeڈ1HlQ XwHgi>p,<±A#[yE2Fxf{)cޙI!F=mdkx+j!bD[:D+DH+~>]U@v?U/F9hf; àmLĕ+ mtorG4 v(TL?,ר!E{RR,a0ޞ6vj}i6 ePlReߋd26u H C l$$yk,2`n8 lƏ`plLU}lFEKE\,)MQ1ƹ关s0I9$@0jը,T-7s lj 4cx2/P*3ψICpSpJ?=UB A!w-XGa[ɪ h IZœQ6H4Ȝ0f`}BOi!rG^g攖S:Y* )()RĹȃ*V(Ɩ-! b)/d`  3y:L}B/(Mqw@䒳'I`V[# k7,n&Tkg(1!`K ƺϫ*w_-6 U~!ryaY:wSJ :'?%S$W!o^~KX^!vjC?;]d:[NwȤ LbsNn_]hנB}qܘe #Ѿ ǀJYkNAIXזxz7BUaXEihxj1bX:eA UV~ +o Lmkҁ/:0I]EsaIAf,iou-.4?}ð̶jJ`R|T'T> *)}Vs;Yׇs66=U1oPHSR#;SyTj|^ܪ_Vn L.>L<ŃF|="_>4Wac»:1%P-T#+G19 xLb|3,*j(E\JV枎9.fCA ]dв+9,+Ǐb1g"&tu8aHʱAKFhKs%g۔JHږTrrV>U%[r«fO3*OTYMCwA7T'^=]ii#0ъ:v5|rRю5VTZ"fXY*jaeТ}~U9W&ᯡ|_RZ?#ro3ȝ3O!D]IȌȥoOfu@AcOl@ma\wLu!rDOUGDF\:> 4ؘio%Aׇ+/Jݚe^{U.\ZbſKX47LTL_Oe]=mtrMYy=&, F: {oәP8zk5w^rwL?֌nN{m ?5,93y|ow>0Pr/w37Y4n59֖+]?pZ ɢwss%e\iz1Ɛ\] [>;$+bHCH9E"bc<q%15 M)%l{ƥVt({60ھӂu|F EݝXŃ󀐭~ye0޸lp+{Dhiq,cjĐxr< ߨwsvߦJK[5ߊ&74<7Y2wZ,8XxX*B"ގLx+oPm3mC qY)RA;rTg-TرpyY\3^q֛o9_bߢjF]S~^YYh\LV-"!: 6?{i\z\-$eF RhSMܬ:xceJƪN,tW} vo+ ; ɬ[p`CctJ:Qla!svV^<./ې {rNKw/fYDj4̹:Sk\;.@-nӄb Wf3}t6g#ddæ /\a{h֕" ť1hk~+`#`L7'`}O IՖIѡ:y8rR=;S' ѱ}7}L: 8:7$Q9v*l)Dhp9vZlٷ =I^0X"߂^+ռ|ZT6F5i3IʓnnQG1N&qBMGҙ< le8I>h|BS=!4' kZ͊Rf'ɤF wF w"/IK R~p0#͊3lvH҂/'$qȉ$ϒyn:IVDpV #^[Ѯ2 ;Eʀ4 :*WīQa5\ͨp /Fcf U|h&G[D(eh"ZT0pu0Z{Rt;CYT) aĢP004.(*BgG"*@P3QQPWF^c[<Om-zm Q+J{]q:ww1FW̨D3UNh(݆tct. K]ޗi%&Z pZ:jϱ9͒|y!<!ֲ4JD <Hl<ҠS g%^y@A;7{?mžMs^3g4+L Y{GAXǸ>dt2.c.-Ti 1ϖ%1t]+MϪ20;54*{1u$j!2Aϥa9 -@t|mtp*ώqF;2H<>,lWϑFsF#-fE/UzBD$(XDR,h%$s vy"%?h& OQ25Cq#0Iz7_@B2-I S*|0nFs*q0L46`x>LcqAbyoqqb^.s*{9#CG9\ל;K_\pqk[\,-.H Lz  z1smc+Ř/  @Z+춈D4 eł^,-.L 9 qܾqqXb|Z$|˫_X.zt^XFQɔERd 8RsZANI`S\_@ljBui@ AOiό%("TI岶䰠Tn+X$o8g[6;f#)" @)`D &Pɪw!9rT,g #3Wʺ< M+EZx?V O:S )3ۓT jC:?J ȮM5Gʹ$Yy؟jQ8skS(EofܡZ2 @ \H Xq:k>,a[ՋK?)8/IDž>,v*5| P:GB]wz#ӂ,, }_IAѲń$f`7s1*}C]K`t =ܪ%pEcdɇcn`ḣcn4`z̽1Sez̔ 6sP{`x1=?P tM] @ݼQ|o/8^^'K"V-g9RX< ؔanC*`R.T#ȝCs0h8I'U寞Q X@XeJ/Ol~DŽ.L0iGAq\w2]m2*@enx'F m:MխY8/ٜco D-il ʭg`Zϋb~] +,xQU cBFL#` 1`T408|ڬGt2[EMpC! ON.tEy^&Ti\`԰QQYvⷐ1V)A(  psh؎<ZUBp 5{(0=xu ;=-gdy,T3`NZ-|Ɂ<NE43"BH `;N2YXk'vW ]",ц2%XXfB`#CF$TY4f" sAT eDbF577Y?p!Ld"F EҐ0Wiԋ]C Y(O"|ct4*z:ͫ"נga0Eb*%!iioG<13r̆/i"Y.G |C縣 E)FDxza_:KԪz~510uV ކQ*њu¯f5M}u|2*4^rKryqQπyW?!]@Q"RnWϸ.OFE/a=oG[\VUֿ317O̶ q=;muIqa?|9YyƝk W>P<#!V1"#%v(`D!| HEN!?7,$ct0@}𑲄ɷ-fLK%($/l,f=UWJeLXZ$UM8 Q*l}JNհ1%% :XB4S$Ae|eK4Jvmxi *TTR!JM <4wc !Ds_7Ƀ %!#)H<{Ժ<^ȝ6ȼ=dqK N[ cu[yu  yG4{f޴!$MמP| t'r}qt ry{`j>I{F#DBxxZ#5ApQ.8㫹(Ƶz5Te~V,npfZ$Z'aiXSb~+fR3V*d#{?]vE~ 1ff3jB,D|f_obvI÷ʏnt:J[i/~mߔn~vP-+qf7n?/:ɽ-f׫UZjoqM9fI\ūuHl 7d<; 7ҏ$LWq +x` U^4^2DVۧ>QK˅Vˢ\r`E-z傭7x8NJ/$g|3I+jϏV"g_$! XLEBpm#a`E PMjWr .AzGn[z=ᖼֵjGy2Lڈg\҂O =uRL~(ޘymPa)k$CL(lk_R U֓d +} rb_1\%cJpU7b<+ 5Q]K%(1!\J8ʕZ3cZ ;ռu8+6贉*S>lʽYr#!`Z2nZ]>.%'a6u~om4a΄x \,8-kpLL,CXGB;\~%"}bv^0a_R`е gk`#n<&j'>qM[vO8Wc)[¼rF\{Ǭ+,\D$&Sl CPDZ@([6*[ù *Nѣ)ez_iAJ;†l^,5Zݖs[aI.I%t&vVy9߹kBHFЅBCAěYFki)~ru8o0[3(a4?ds'@q(  +]O3 }qSg-w).rOW~  n滶D$'ҾÉ\4;[l>2713 ;>`«2% IZ.Њyvw/k;F>>^k\ 5u#qӛpxqӐ_} HK?U=wy;Wuy*_+son+H*5$gznS_n .߇bH=okcN]5ۮ&PXޮ3g'nن+-{^g ,[#\3oͲڵ'^T_֋sfCOwY@z$c0ErI+͋YrV2)Ֆv^DMrA&0C]Hq'J)yP\ӢZHx5KB%ϦxyǦȦȦȦ*Vi{njPzVHK%0RZ4w*s୓ROV$/Ur/hidleMjJmK4H,Ml51 ffCeF[Fg/«Ő~Qn>_E{v* h P)X2`>ޔNSBbm"t&&4pk Ƣ [`N f(/" 1mF\ѣ)m&ާ8h)DIӠOt =MbKfׇ RKmNe&5E=)4jNHښ!c94")9x]I|bH> ̍GJZ9M<˷!g]Y4Z zg &-^+QWhI2oz9sLIy'[U۠6oڐȕ96]AmfKzHص_lvlI03u=Zl3~FznG +}~Tnr+i*yLHp6vV>ds#(laE Rjan(]ArTJv|Ol>N8?xhr7F\$-K w=eWTṟO(B.0yK=x'oiV4J0um5<L?ZAp6YhI E)sr+'+)%tj"p=$L,JsTl2 c!3{GbK{/&W;3p*{>gwvvVVe@&e~]¥)}m<+[dZKQhU@}YOHSŹNwOveymJӧ^smעݒ=-ğJOB1F 3k#QuZ'4q,<*Lxf`:tV;=#WE2g"O!\TM5z&-Pz8)d^nW,*)L@,sH;t&!-gC´;UI: 6Ue1{#`8(a=iM :` |'CKH~vbU<$o]vՁvP%i ~m%YsDqd'S!sk߆gWK:Yz;QWtV"M0V?0V4}~}T2mاdn<]9\oDoz{4=*;5`Pay3[GV'J1aH`*tY|TJؿw|% Y}3x}u|Yh{X^mTluLF@)3}cJxm\⦲~. A2=c幮~L1I1_P)M8Kj1^iٍ1e,\x@u&;(Q]?_:RL#51%XWArM+ NN_ r)tH?Q Cw9Z0R՗0c}N)%#Ɣ:@v^19nFȘDf P_6wtgKW; (/UC#\%ha{鬔ܶ0 Ϧ#θLqF2eZ5.Վ@y"NaR0 PCv8KrNflz%D\<}M4Oc4Oe(1p(rW9/8-QS/+DyR_)w5^ʐ%UvZhAUO'cq 1hK$g4MCQB2&Z,"cHpE*рjMEh.y9 uho(Z![3 V`!̈́ 1 R`Ւ`Do@$\@4ByUK+^xs`)FDR ) yL~>e;m}iݦwV}jBYd "+-MSKٹ G2k9Ma%gm'jU vX1-t9>=őOy-oo@@NA?W,Y3a\~J״fQr&>R.{ 6Gn̕Hn>x ޣDU:[  ^2gf6ç[ސV$;y=J1_}72UGn[Q| =Vt2 Ur $FriЀS jw\աķzkGTV),_9$ƅϸar&t 9иx' LƉuh4? )ub՜xOܥNBO{"ϧFy$OcRfp=cHCw"υ6`|{J%϶7#hR6UJ?gZi|#t>OK_(-^y̆6y]Ҏ/)?yʅ#Ͽgu !]G\~[WG5Z49[,~QտƟҬ犐?ߙ?Mnbߏ [[Cjn; P rc;2֌ G0T"b{DQ@~'P< t5(u@ AqGr؄ OTx ~P#uzin P jJ ,4i, E3̕hE>3fR=oʿ\@']7{sݮLEkTոQ9j<(}PgDثK6u vjk~g' we 3n9aSAtYQ`3"r#~ZQ&p'WrG֌Fq9iMn;@Q RPN1_B7FL}3(Mn ʛ! P:F~Քh[;&-89d>|Ņ(_[$X0Fq):'Zk0R1XC}dAӎ",0"GAQT gI\/_Q/Ԣ܉ . 6YcO :eLi=C@ 1V@@E:)O%Du325FJ4$2 D(+E/=***^@}I9- E [O1 G0D91!E4mBN ˸ V qK%XR ,T 2$Fjɓ$GP~<fR 3z1Xe2]Qϭbix&cQ.YV^}x|s*fGL_[vdD ѷí:#y=v B8)?|߼@'@:X@'l^&)7h~sׯ}(|o7߹ˎqRU}JܕPEdOrN!$^rt&L2 x,HṈ"Ғؕ#):ң6\TGvps2 ͐#G26`Դ> =l0))بgW{QԺ'G;(# ]?q8jҡ30syB|4} ~m$SF>0!\*%+T8ܿ,PR5B:E.6&]^S$ ښnCi%WL )AA"ބrN%I+f/rd[ e1F!u"Mĭ;aF΄!5 N|mBbcbC"a8p ]?^e#dp=>7b!i?v{j%|ԡ$oƃ5JXzP< d@reDA̎Fytמ0Y'sN2Z2Il= aT̋ŗ82 sl^`Ǩ!6 :C}9ԩިQ12ƣ,zv AsJDC*Q-v}y6rS['smnH/yQnTx\Sb;H q@Jb>$ zw 7@n݁C^R ^ J][wmߵUeW1K(MTkdqDALRu "K3BG1&'4Npꋴv"*o/?XU+7Dyvtk뿏{t D d7׌*U/?%)(%*E!, eq 1yd"4W4*Q8&ʢ"qLJ:3*ӊْ4Ќ1âTqc/g ÁVjGXiIbRd0i6FgX($YJG T# /'vH3Y#l.ՊXs`1dc0 )%( (KOn!An4ĜҒ 譮wlnm*| hhv^=Ŗ5.Fte/0"Tql̯nvK8x+81;ݢ&]uvacm'W\nx,5jA5yw?D>1̻s"Ju/|o;XS[#Թ8p#Iegh{vB' 5UA] R,'kqkFZßL4{32ujsp_Yw&)8UE.1v=0a2a.X|Y?bF{xҋyJc޶)eG{gH'-Q'{g6 Gbipq}|e_Vz1am jyٸ8J 9%ٻ9IIc,hw=eGc=3UK6tn̢ʢBKEpDhD) SSGno:ؑRlTX5/.Gne&WTwKS[ !u A 0mY]~[ )9$W9 3/.1N1F2b.]!ҥZ'BeW&GyC=;1$^AcM'T[! j(ϡzSe1"#TI TKTr#b4$&aZv0='ɷ,kmlu /{ ?AyoW"C5`|\Y}?9Lpد~hBGDZsFO%bɎwyݧ]B"upR!h&hXS,%_pQt e{84HJ^ s4(Jus!jB[}P{O>hͿc hV(wyCP_: jm!k׫܌ٷ<&;DhE罟Nqz_;0l 2XnjI Kn ^6[JJ^zu1xo󅹽XbQ^j(.&\VL%|lW[Хu8#!kl0pbO6jEaaB@|:0fW5ES=w8磛-{ |B{tsJ{~Cw_c?ݾ-/]`Aɽ wzqOt\*DOg\š dOat%Ym%sM66z#(֭R~ ^x(sR(W<|p)RQ Q .[bOvNA<!u~!;aG3^Uw/%$J%өu?$Mo_S/Ɲe?Sh$߸ֺ!tX]{O19hA[ikS 1(XGC2"QJU*渫C,%t㨦)٩,gK81/ y+ukgN,mhYv5Mzw֧R1+ʳc@m7;syv6t]q1TKb±6Zq,M&q*n$nS3').%ʚ˃X4RG:?Hl2 XIډo)~ )-TΖ-!Hz> L(MSR7diC iE5ز=fxJQq'$ vNܬyNj'wO1Jp}^0LI`a/DŽz0}:qWaOyV{8wnW#g 1BO1o9\R0Ȏ%}#w+5߯K$ag6<3^<6_5m]5m]erd6 '&9)vsgd11BiJJ0Mt-ʖ an/ EebVֿ[\|2ՂKDDʧɬ >2]oxtնJ=-P8&Drm .:iX°L:ѱibJ%poyo!vxow.}1od%K.fqZ x Z"$'Zj$ƊPKDH#E6#JAP`!"V"mܲ &DRHjx(լPXӠy1 s`X#Mfe C 0ZK!Spr(0|Ů4rb]N @$`F4b5?UQtDAФRK9&K$jI5Z ߱C{.aA6v3:*r0(%1hE)KmL{m{U[W{̑`. $o*.+4JVBӱ%t3A3RȶаvB !m'4 ;N[\`*aJݶH7z15#*2TRa4lK)vpՍڀXE/XM„^h4+8`,fg5NIJJ$avR8W[4RJ]%g]*dx,,{42w<"P9S|&q=RxlYE}K+Lic sevl%t!p5+S:jR_fuEZ&R_徶cT2(̃}E1\`G\4_o23NJw̦KMԯ08.^]7&a6a[zUВ;+S8[{XK($YſU~ ?Vxh3`|XhWk _ga|ٮٮ^  lk/xmUͱ.NsJe_'Ml[iY  YiVa!>r;bj]W SPvC ̑l6+M df\0HUx*!@Z$$UVx]]X*>.W{t6B]q% |/ Je_TPHWvĂw<~ܳż0T0(qLD<}W843$T.vܣLh,ky,rqߺ*'ռYP 7%mQZS1/ 쫁b,L%ci,|NAlRA0u}܂Y8 IH]+Fn}ԈXUk2O͘xK޵RJZTe)4GP$C%a{! 9[bQ>O?Ta@!nG +8]V+3>T\K%VE͏0#t;~X)^=1Ʌ傃s,V̪d* #t zۺIIrѳ!#^Jmڳ``YjQ"g `My`hF;О7,4>_*SYO1dNIUZ,d  1i'dV&;{$Vn 0p=K:$U#b ,y($F*EYZ$H1)ezVb׽`iOv{?iwA52*5=LuQ'P:?&3%vӚysX;@ rĚaN(lwO43_Vj c|XHŽBXRz!|~_M?a ^_!aj3~fC!xզ0"W.mDZjHG)} [s{|(hƘBs ;N4]WzZKT\\po07CckNϕC3in5c/~z Zx}a7l-;sC47y ;7d;7X sGL3^w;7m^ k#`:`gl8Ujt-͓\:\y'kL-ʠ36IJW^Czx9w76~0E \.JrQz2A_ 9GEc\ug)\)It9\q`cE3liV 2t}`>S24_ЀͥcGc_OppPA|WwSq59}Nhiz5n`|cvKRr W59r6( Vc-}ý!kʐO$]I+zQ͎]L|o:kk̾ll_G֧>y㦮ܧ:n5b?Y hl$裃iCsȁG)2KTdJɂ ʹ5 Ƕ= k @hy@<ЌήQQBj~,"ԆƖDL[ՐjT |!s `AW{2;g2&I$%LT'^ WV^=U*9FmJ"BtL4!F1^kerNxPPD (x}sW'wI(nO'Ո:m.-{: oIA  zG%wtZEy0>b"#V2aIf!Zw>Ϡ|Et1otQI|Q2O2{_3V(̯\P􅊇?~]ΡLS .{ o_vR״GLkf"}t2\]^'OS>~U?\Ɠ߹퇻Ʉ/x0͙^ݜd=充 ["P \6K[:TE&&{Y'9 5̪qBoGV$`" ƊAe-VTE+ fBLU\b"![ ge}Cƹ1iW__tκ@q7 ތ{IGHg ܢ/RV,BR%e 9wdM*N4\^vI8j8evzR[olrl $5Hkc fXcj,!c\cV܎ϸ 3ň+)ɏ+%B'Ģz[T]sO 4˃@4J<C`NB&} :Gz 4䴷C@*$hIAQցkCO~{%f3oÒB#$x(>25ө6Yёc=#rZfq@SBG ifThe"q(01$m%$yچ˥b+c7ɂZl٩b(x3D\b#Cm#$S F8;UG\Ӻ]YLȮ=7F^j*iޛ jM|ò!ʔ p}KvtndC1lHaR~̟lQѲ~ ö ?JZsqER)|?h;YrtÑن7`²jp!K:7utPV2W>rw27xOh*ݧƝ\xO6l{?|W/ۅC?]9lMxeejj&Z!c&v[r &vd/;ib_L|3biyoկ]xnno ,.njT0 o h Wւf#LXZF&E҂Bkf&5ͱ\ًh %c f䂴z{ѓZRe~ }@2SRU<8ϻu An+@C?0/R 4l9'z.9ꍮXLF9+7Wz\27(A⩩4-JX,:z!KI8~o,uw\fҳh_S,X9+ڑEt{J1xMm@dD6(⇷/sQ-A#R躘%I2\erESxuswVص~8?9ݜ^ѧE%NyT7UVFaRVJݞT ~ ]R0C>h>@աx2M7M;~|H؅đv!.%.]yPiF{R&)Mwotz}vzAόt&ŏcw9+*keMwE*^塿ϻ݂B*C8.O),ZC%9C#l( Rs`HwәGN9k_@=ג׾SN.[ﴈ9*[D%O$'H| "Lp=5NS]FϏ뻋Z!@ek()<*qNVHeէbb8O?4 x.u pe ] u%m'z9#Sv+kG?vm0>іp`+:$鹢zyH)OnC ;ZF?CIN7r֫u<ك1b=(3vv#[٪S}jQ5Ԅqd| 8}8}U_sJxo<E5O;Zqh*=jlW恸+L%So/CF1)no K[BDC)wyuh'&t:"4#2@h }2weRs>WܙtsfKʑ(MDEOw8t"@r`Kރɩ(L;Y9S۪hN5MiΨ[4٧֒8 iϒyƭkV?8D:Oʗro.::c9#B3ngv7!~KWN4v=#w:. AGQ=krNcZ8CپZ-l/@ |b ,`tF#9S%x$2Qy@2ыP 坲w(\ހk*H"s:5U(Ѐ' IoB'aˋ<Փz^tS'O]°Br.36*ap !wR-]0Э2#!ɨSIz~9R DRUJ>˛JMRnNa_~MX#i;"'M|z5!@TjS=ViM560: aayX\ͬuXqQfqiB#Ҳ=@A71ʬ9pglJ8YVuR]r ECd|top1*߹-ۻ8m"+zgV~t5 ,(j}c"-GKLQSJ&މGȀT2yG>-oܴxt̩r9ꥌgkA\e wC/N(?)lqH)8}JkIwV)Ūq {ID:ւP [[qPȤoT"zw H&pكNkw(~jk)pmN BuGUJ57!^+a*QsNFl-#u\Fą:*t<&2=QJ'j q<"&l!;kriCnIm %0q-XP*NXQG$1$΂"!!F%@ 2vGd (#H? !@sR*9˴t>qim|7BG"5QjB`]2qMഐkEksB\H4 dTF6. Jl.i нR-7ھ&]X먒I{!EɹjnNfNW12}.W ^0$^,8xЁhʣws|tں'vun@mcRsF{louͲ:yn 饜K/ 퓡mn痳![Njem# g `zy~~ %CW"n<++!)io8Uz5v3A4xh(M"K__N y GsMRXjW/ph"z ~ /-#D`!c뿾8` xXT$B_^No:$k@5Ůx储<1Ɔޣ4]pq"H^zqםڥeT3U Pd` k[GNǎ1^B]&p% gxc5QwzQhY\pʤ_Ÿ-{mޢ$P\#ܞ\~w@j6D +u9?O|fE W5'0 W4%nSJjp`Ӈw7F<4z7A((wfA7Ya3^ONv {n:jHdhd-v<4doz}} ^Xߘyd41r9ү0h6N"1 6fpߢj7qtt g 7\{ú\C㵤')X u[9'c;H[Q0ChYw, .Kc@aUw_/{E_1ƟQ-OŅ^zШN2 ,W E8M?rN ǝ zZ/ DZo4󞴵D}[nC6`ppu<Z3u J%(w)~i)Z 1^5^(8p-w}E$-[\!;(H(t)*;.~Bnk  FgI( ȃ($A<&}jѻ:&s  ǀ-\꩜εocC Ĥ2cPMg: y6ǻ<14^=CpV1yXf\>j57>m 7nx~LHӠܵ,S锭 96 Mgk "I5 F~˦7o/,MmiF#7s|Pv6ɾ%E'3|CNL FP RUZbU+D1Fm.s8 NKH\qWB0rjyR:F- -WcC1p\!HZq2@cJ^3=cvCõWE&IZ*5*iMCcގaOk@4 +rjE}\ _B%ᣖED$$@P ~}_l ӽ8)?Ȍ&{{QGn֛u=F8yȸWt?MFav a~N8*xqǃdPdU6ppXCh8j!EksZ܉C\zs}ø-BwfDnkl g #5M *GgVngVBl"|j ܱ!͡Ґ("ɾYXK-3ӛe?gJH283 D9~XZ:Y1Yڦ6DJ{Zf&i\zw?Y2&s^p'-hgDH Gĥ œ>}w :FX(li7`/WB3u5i-ChϘ'Wn{5&%#1GyoZ 5leyVD:@)Ґx>rhB=*% +BdS% c&y X#m~;Leb#XnN1U$e9̀9g,}+H"x9- .:unfUt~o :PG=3EC S֎7iWȢt*^Z (^b$*ҙ|Ve)K񋒚Vi\$,x 8`cmk @aOnidIey{M1"v\ZR=jJ|[Igl#$G}<836R Lg+ম^yee-v!\nDREhF4]Yhf6B6I]D^}9瞧,zt^H5\ZjɽLׂ!8َ1^i2!E(j==VHk,\oAq~M:#&뵼bIc~В6μ#;B2[ ӫb>Rӱ#KQέqlj GI/iy*?ܜ˧kzK2XmiՊ-arL3G_L~_mwq]F!3|OZ WA ]P xdRZN(|]w%hk.,`Ud>W^mQc}rvҧeX4R*ۄ$r-b;x=ǘ oZZֻ%E~,cJ)u ףז _Ih?;yϏkϏ?1]@Ε!y9*4GzZ^tM"u[Qeהr U%4MGeAhA˷G>˃D;YTuJtqi$MaH ["ic Ö;l: '#ttGktٮ-e!L\ S8$*Y=TMR>݋I+J]\8glx?YߝQ˪%.-9#?{zsxdq,(ZZO/:Ltc7[OXO(FE{&4Uho! ~vnt$btھos D!iDQ!D w ;Bw,2,>P Iɣ,!X8D|'BR48#-j1+RkhYV5A#Re& n03H;OJHmku;1%8bc݆jE6CV3WkT.ڽC]Y* C R0GKۄH5Ka kx=KZii]+-kuҺT)SNشDS,%E)'!X9X*Q4y6TKm|i8*//0aBU"oIZ &n$^n\Sj"vTT7wOIe׃o5V'pT!([dK5sS:ecFn &l(hbBܟ[A?7?t@l~?L'I'FG`6UM-w]m7^77NZM4WJKBb`fx6i469-tApT~X i_6Y\^sj^s0$t-<µKםO=Un} ';fp tI>{mtQtew/9߲vx).n|߅W HoӰlab|4>m۷-qc;gzEeLYgk1"*gw;S'S>S?3`jWM ixO=]A݋Ah=u1pxꃨ}<BLg}ߓo0q[(mQoNINz;$ (5nó^YLJ|~Tt亂rIX'nUeÿ t%F>;6 k0F5p 8Ds@q,4PRxrpfcw+pq+(d7% b?nzv_Vu%/-xh?`ЭRBVJ*'t+p5up)VFif;G=`(%&p-!c+jZP-?nuC2hcZQR9V*G*(K co]dFQexD0Sg14 [%mtm|`d/MIbS"TIߔ ͥ|Ԯ3IxfY[)+~+eorV2! XE` Q2m$ :m{` G`+ BCu_ڰpB+OP1d>[ud'gx*P,"pV H*uCU߈gY;,7$$0Gf2}F˦B E3?s#*T%h:̼TυT(4&_~3Œg_k(g0sxϊML(ZŽo ɗ-ZjP_ !Lxʛb'N^|>zJ4ó쿞n'榽09~|blxufs?crgE/f8=\Y?\`/}cƯ0'ᩗߕmwr 'מdtf=U*.vLE 1SX gbzuWUz?Q9s -(, G$YJJ 4u!Fe46DSkxy/^^?Bx{ۋ쫣5~@`DPVhmqdVRƃ@<:i h= (x3- 37iI"T3 W!QR8L5\k4Fk4Fk4xhpZ'  K  +/8x @X^Ё޲EM (a \a@NtܲG>J >)4x.k4Fk4Fk4xhpܚi+ɜ.V$ zynLƂg`ʝrѻuGu_fռ{0m-8d \ZLc㚡U4; 8NWV yTIxL31䁽V*p3`j" ޕCc-Mz]QA؀8ZaZUw|Ul2g>RfvHP|: 9:*I )fb,i ]+{}]0wJ3s :ę6fPY#Hf7r "Ui*aDq` Z !Q`wpZqV몾(BH$몾?N@RеGzs*$~GIO#ɞ#PZ1zOi$ץGOFK \boΉzI,0,Wu1FM.^mΣT3kbF0Rɕ "Fp%VU5 E$Q (A%M[%f7a'y:hG5g_"X7h% pZ8(#jXF[R>"KElH K*a]5 R(S)"cAz!qL::dHҤVepz 5Q1a12 ȓΝfp)dn kꢬ`,'hS-=гD8nI\?;pTa(KTa(+Wj䩣RԔf~c޳E,$Mb2 DWa.p.a|.ʍ !(Ś!Όb> d8Lj9OeKB`qtwDdɬw?g>mf4, Q%ĔC4NZrTb#  "fH< @ԹA ]4W[J_.oidrY$vj )/<~pϯU!7p ,f C6_!9/78yYЮ<6_PLI %ʨ}(0V>P  Ȫh1^CcB{D_j @φy;19Q6 񗍁x4Tiak~x6v4:%<%S7[#uN,:=.& 7|QEONTv,[>fJBWL _wSu૱+mszU}C WوSk@0Q5h@(/ {UX*"]e_iB"0!L-ܴmOCn.aW۝k{tGu@V{YcRv=&+B{r~[c) !ecbՒ8k:9aCZ0,iCw&^F3kգtٴfWEy!˜GjIxbB0t 6BH m+K|=`j9Y*L#aiQ9g,"Q! ^@j9@ԘuLK!ʲ1szպz뻥 ~;"J/GD͘9vZ`=e@ܒ0-G>r"+p`Z@2"c$غ$PI6jwqZGYx٫xwdm->hj]C,"[2dfƷ&p%N*S$ۤMDp0TfsM^D7dqEsMrLUH1 7%EXWJD3̐1{{Pej.vzB` ٰzEW8 B]k~{KjÇ{MlJj௝l\D5{Pw)4c6> _u0{m>SŒ| yuMcOq$;g( y$ZH)]塚-׸@Ap0B:uXݟ)qƸ[Vn'Nuh Nxu{ܭ. N1nR8%bcܭBrUP'ѥi3Zvޟ,!5a$ @`aH:Cnꈲ&82?i,b8kgWaʕKդt:[y YR>^UZ:* ^\8laE|ش{*YnL|}WdۚBBmk>f(&B3Eꊾ-6R'WMQdBkEFR (YޖS\>I+J'ZϧN%>oK0U'(Z9D ?]<`G uCѰ,7gfӶϊŘJ)YϊgŵxG!DE DcZl]=!Ʋs xX"Am#VH3Pz}` E Rh\8RBaR2,l>A_U?{"z8fNIx0SRa˭lqHtFB3gs(1cK%^kzh W2Džgz6rL&Q>wܴ!& Mt4vl~Fp1r镖cp66 YOe~6٨6ykt47=M%'vF1ͷM1rY[7"AHφ1f4cI/Ml_rꊛy&[8N55F10 \݆\2%xbx c'j(JX)IquFyx$`fgiW VeG%uU\AuT(,/PNk$oX iLm/ͫqUúw}cmH3$X,:)J&\biIR@9QNRB+!tOT![:WnKӉ"7+|sK&+XE\hxR^DR B-R7B\-])W wπ@ٻPM7=@&>S3,`o_JM7)/~Jnv۝ݬ;ǭ׭?zUCz߿i<:,෾zW~'; í<8y:~{[wA3 = /_z=]YqyROC_ހO`_vN^~ lpBI1ii\Ėq`o_4 f{a:)xe)wd>ؠt{YYʥev` u;,fcf`;p( ͺ2]7vOlV72G>qpg!(H[)2kMedfȚ?32_FUeyko8jSf|/lGHN28=xssa87Ŭ_v#0C\D61} |ɓ\Ö;9˰8i/9}DwF?T7䋩R8&Bc{pnlwB jip-Cgo {ozhL_0č=8olQ$SnvOL͠}:ȇz}=gYwؑӻ׽q}_prV} :w?sƛb;ǵ\?@ 3źާ>KdϾd ]D/߾oq!W'ҚqW+Ob~oBq%\~~ f.[HQ0̅isxʙs`5yI@0+q˫]y4@j2 Nb0F@r{ l@D..EXĮh"- f'8cIm r 7lCvVE!Y"Bv3.#י" ,5<}z!W$#$RބD2џc "#SY?l 2Ϋ{k083G<ʯ"R,򠑫ӏ(;70)CV)DujCVISH;""NAZPF <PAXkִݚOɴ%?˭i5mִݚ3mh's`nӘn}o2׮Z;N6az/NfSGvQ#яo.S# N5/r;q@BC5 LȃZyJ0OB"e42F*R!*slEǩkv ?!3/W}T '[ 8txX˃U+Ѝ(bz; HrO)PgƲHr6(ۭ9;$1@yhdNTؘIH:$b\Ljs (:|H TJh ʊR4FjxG- _ 5$d"{yۑ5 mYHhlX=rG˅Zތ4 9 /ZWtq(#1x ^=k  z~ؚ}T@aE>!xI5Jbl8!pRp(䤥'd>Pl+Hww]dYׁ嬃) h>IIui3'ײB-]c&36m#tގ=Ad;ZmDsp΀20(Šf}onw)(3j(R@{$vFpw#k" 6ykw}Jr`ũUkl;ýZ5XI1{&5$d.yc=b]^L+8Ar+'ԃNj@exB:X2};+p^g[n\J;[s@F{LpN"-faloKiB&c6]e'CHgF cd:h.MsGfE'~7ե׼mߗFlF g};7t^`o?hSf6? 䫳ˡPP$)ÛUtZiBӘb096ÑvHMv*5*s>ɏz yVN3On׭n&xM+40['Өt|Ih=2 í:(֐0Xd. R5K"P\@`SnvNwr׫YY"zB\M˱! )nr$W`ݐ۱5 BO?L">ԜEףOͲYsF,D # G#V"l\>؄z]Kv@iZKLMf\ʖmWmΤqd v5a"-W#mE}tYLȬ<;S@j5ҖL"K)Y˦х&8͸r>e[ٵ i{V#nɴ.ߤܟI8 d9UH]EOJNVHxA쵩nmHJ5Au0T~==Hg]7~# o.n>Eyw{v_V_ylJz|[dʗJ>g?ɷk 42vۀrJ{ҷ/Rȅm AR/oZDQzs"H |q\gg7EԫE"A""OˆHTKkt ݜeB!%a_m#( ٿ=!ͼß,^҂{dɂfdGЎzu~^}$;yoR4Xh!j΁R3>O9(S{e aqYםWLg9&u(bK.}#X8{DđLIgZdx:֝1_u_r/ x_T}C0]Z$iߚ1;@L<r+j7{ FWq wn^zSIRJY)hII-9H)a+sAL r) :rvex9!SxY*{GvTlK\6*9E%itqF䩚Vk ’T &ͣcFl ڨ?ɩ ,Cϓ\ ٪IEoZ1Cf&!ԚA&q) gvdeFwmX~Pb O`3;fwJ˒[L%KTJE*@KSGWԼO#!3~Q0%cFXRK a$Dk)%M"&L 'iLցs,Bliб+t<VͽsrSb3{:Ud̅Pb]S ƴ=68&ƌ3̧FۇӒv qRfjЖ=Ӽ25<7fhD59|WPfD=zzw`}WGyG\m ]Xؾ@ L"y2*ʲKhO5au$=7;J:2L޿o2ـq6bmmy >@qllTh^zrzg0ބW)*x@L^<Wn޾:3NA~%v+aV* ^Z eENHh}{H7ՋY' |^?<edw5nxplo| $gl1`p|(̕q8*8i;NryY<,r7O9\A 0 =1+RE 7m.· iӍ_ԒmKTA^$=)^|?y9z*6G$]j0ff'ElL{z0`!68 `m@9ч1VAPI֎yBB0Q[ߏRے;`.l8bꨡ9? a|:zOZZ0w9j)]e*R5(y+PBDkVɩJZIKH<:!Be緑.q}Mxc1yu/0O6e646FT6~Z^77F:QBa |nNjU-1Y'sg_㧸\BuwG90y@#Ƽfa {Q$I/<q/l,:[΂+;]\Q?FLj80[d9H45ǭ t t|n]}X#^!fzE/٦vS3':Q@ZݟƤ!tڨ"MXG̏4PN4X*@bLHB$>lU1.Ͼ8cm h{lmTݠO+6>_OD9ct|P`~݀i"0taspdqq(3جp/vjQg!u{\9Qj?yM!WwqFaA8q(6E,D2t~FxuwSjW?edú<&k|AtA.WifxZ`~~[9dW'r>W b2b& &D1Iy82a0m_')FJ '2:G>??Fm``_@wM-5xXiM!= $?q1nBuT:zXә@o??拧{76٪gᛛNз͝r!o7cyy9/urw-Bi> %$qI&PqO2(x ̈́&2M4fC2 94=vP=&8 ޡHpr8ϱ5Vc--^ bR)Ny9j>בbC/@?m$[|s-iiC dJ @$2]Y‘!A޻k'.^Lt  4Z❘91F"PIlMRD-KцAOԕ=»XJ7E,QE^ ;#THP'TrU(AAr]J )Dp4/ %#J68 d@*##¯/yzD:2:eI8Bs ҈J'(ӇHP1Tw9ӬKT0a}N\ںCem_$3j*p'*"HK#=*_xBJ#J?2yG6 t1dL1rЁ b[Hzr,tMV_6%5a?RQJ[9J$6>dZ[ I -ītTI$b5 :$21ߑ(ND$ X f1qZ0][j&zp  7H0ۥ{X(~h]\ݛ9U[tɛ#htwk1xunqb.QfRQwx!Y7=C&7@\ R o ك# kJ( S~>G(3ř$<@9Cwl?u}TZG*yE󷷏~ LU`;!/]2}pb1 5m84VC1&-"-(ġܴo(K|y.88qLaߊI1t|p?LݯR̥;ѐ!H(BPu(̡%R:VP,lQHJ' {5M^Cj&a<4|kziF`(q#rdF#*'ֈ)mf*xh6 OI8J+!\6[%bݩ ls보2SD+@i€Ԋic$ara &][dn'._uWNp⼻TpEcGa;Ոt$nyʑ۞~ڭ Js$5p#S։Nj#H5v{n!B9jW) ^l/Orb.#)E4waC֦` [@ `4v%:5Ծ΋\veV O٘Y1kY!\X)iSm~vhy=Ze#sو`PY[2= 9}Q;L ڲ3¸ bbrZhu 9LIET+Fc{~9kCin8,tN਼BJ@o5 rW~Ec$IR+߁{=53$K ;J"HBW-LP9~;A#V۟Cn ]ZIX‚=s8xPGfŰ&~j.i" tMfib "1i )9 5i>7jw DT `y4HJD(~ۅ^avynUæl~hD"Ǝ{V$зZg_y)AQw?mct& G9~z` īS'5~|v@l:?;T seq 9A6S)l"IG@M?! eF x50XWK* qN1 8e8Ӑ%sy% p:*G1%=~oLʳSB% U^)R] j?'L28xp !b|ퟃآ׏$1Q#HܓQyQHz w %PғN* {H5%k$vӧe{i-% a}"ⴂ$lןFyPclsz8]ɃkY)l4wx+p2# 㾸QjbCGt 17zիamvG sk/1^Nq#* sb3D:'%'2I ?)H4)6O0wdBh'KdJ"8&4FIhB3̒H&2a)I,#я}~9܍>.fzeY2ݏqݍ~XrdmVKFi>ZcVia׉9YOڃ{:3RBg2"L. GZ8v: ݂ 2ڗe81tn3g+ݽ21fJ1R3@ 9d% 9|6+2!)  nj.=SSxA J·>sJL]/qʄiXI '9-KwY;.c;Qr% Z$#{˵UWx3o-#!rgfEv^+Ћ+ ~I~an[|$K ?wGcz$a2[>`&U .f~7w7o9`s:úc/7h> 5_>?jǘ xYd8-ƾg6%4٧eCFz_ں8+P`\fCծ^]#Yxl 7\1.n DcOEfuycϱlY>ckaъ>|Fp{6L1(Y;R{#Iڒ޳3/="*1”"nq.!*h^Y eKBö.&r ae1Σj EP#Q-DMX+H)bIp.Eu塖 @);y" /ABR&+冷z‰S$ d @3f""Km+ V\[yQI~w]yK!I[N_itxr|;uzblÃc[C V⬜YW,)@yBD be%hd% U%J@C]e[qU6t]BN[0ӥzD-"ĥ-+A,I`ΟQ*^.N[Wɝ#A]w-mZ"Y 4-MCҍ,9zu=lEFm5wÏs8gb5y5QRߋK-9yL+*l 9 $,c9g %WrTgP#dd)wnXJ\&,wGXFZQ!f BjQ"1]Nyba3G\‚NK /#,oG$qC,+2C(VeA!mwu)яp#k 7;˸ӯ~0+F7X~^'F.Ug}o: !Jb FrFFem[;x8:ah[3Q,Fъ5ܲk6?1R}Fª6W-յj wK'_L6néGNt cfyn*\.iF\v{78#qV1-l{O' / 2))a϶%T ^'[,Rl;~vo,|ԛ17xEKP+7CsE\1Ox0}{$=(K )d;ɤ% ys ȗØi4|bƃ^H{7^4^n!)v;gv׿22YvhCa alYwAG~@4QiG"i!,YVPO邜YD2n5"cH0Jm`UV x`4{ Zv_O?ẃS3H6&/~~8swo?{/K0~,p1:oG'p R#`軉qpnpe#N{x" ^rEfw[^/7.,o ![\WŃ_:+។-l 8-ZBp x,d+9jZMBg3fdC:Y([Lysw{e>#/zyw@C<ǻٽC3l5~s(q>_9yYf_Z,pk{eS(׃[( ]V~ӫq_n u;aw3u!O< 'G#ޙ9ɚrXnKiYjwPA|rN;g<7=cF;؛YOLĮ ݷp gWc˴}v7ߐ s.{Û{Ѵ;6N;l @T?T`^ ]̅Wh:5Ϯ^:v᫛||!4hqO8x~w686Ao꯽^o)5N%܎tzr 3^?˒i0AQ&hAo@(_}5V#*+j6~JöLuy?w?+/b2n:ghXO'AB@hfk\ğ,: jd Tib`|pMJjRm1gcS9~Co bl0W#-L[ p⑶.P54(L4f]7]2V5f( ``~8*q翇Y!L+mW9/B* B8B [)9-NiQ9X㩿[zEӞ'}qf< BȄn=NReBVݝXW!$Q> wcN [ÎIc';s*It8X E#+AtoM5맶l> n Ԝsb l0Dy8(hi ;7/0SXZO HACY @(C@C FI=N>n_y@\]y D3d ~Xz*Xh&uM E5]хEy`4 EMI VK@⚐i e{!%B*i1!ذj  E {t?N=M(1%NHb\}jra-\ Ho}1E)cJ Nj3Z)hQJMvMHHH"+ZEJT 5`Er!%`j?N5A,B}gsX\~"ۃ"~Ws`wsĜ ~71%׃l"@")YI)s1dSYɑo ؆4}rA;qt9.G] QjMp ɼbIpR̈́tN,Hs[zjFFf^|Lx-:ISOx̝Oyin\3 ~-II6ΰ9v+RӶ{w~FsEsolTWFP} PkϗK Hg*ıqqIRΐŌW8ޤZSF$mao* Yeo(7"*hc#ZFDqP'ZB qи=S-Z<"&k<E)" KĠTU *J@1*_ RHKѭI Qf jL Zzob޳6r$W}K]U (aCX*(EϧzHI| ΐ"%ڀw-R=5U~p)hB]"֦vU] BJq_)M:;aL )/(R(DϢ3[ KLAEIba\ʝui煱ee6"'BAhy`Nk{LuDkJ+*bl!H ljr,؆P=F뢘MJ (IkDS_'Gw:T Y;\vv;+ھeJYOC#fBndhd8J2Yk"JbdWt.r+M-؅c4Kd~IfDMB&)%ߕmVrWS9{/D{ rIQ+VPQh"K.r;H"D:`Tu'c~e!DXpeі!LdJR"Z,@ oAX6([6 ȧ扣OE(=τs6|FeX}ԥ%|$*+.;aLg=kP#+8)šf>֔Zl>PBx$" *BWJ2i?sECIa!Κl}v2'fe.k@(HN[30T.6 SrEJ]8ӠiJj1#݁1Vk2@:/NPzzEI.,f!v¼3jgEuk2G "Kq&iV9&g@މ_PY~&? y)@J+y5ꆘ#Dё&5eI^Ddgt Ȋ,䮅"zL)fmZ‰a(V15Ylv~nEpASFNrшYX`=b-\c6Vl'c61tI ilfDll,fS јXS%cPCR3üt-\?ETUdZYHFcyzw:NޟLƈO4E  R+?UWć S×'/P̸:Gk.rVqS^&K7^ؒҭ+Ngn/g{ܛt;tC8nS6U7rϓK\bdwkv!^^*‘2`5]c.F6lA(a¡/:/A{X:b=m0FmǕbٱJC3R R!Q!oW3bNC#>ZqfbX(,Һ<Ş8τPp9څ2䜦^jņ9CGݻf?o1k66~Loϫi`'wY6 7wՍ`t1f% c>Fމϒ{xו6JIs͞)/OtbC`Is{3v8gxXXSMlCi4r9y*V(0erUc=۠ „.h2tȅ6?fgzta!'ݤ (Tޟ{H npa6`F| k*C9Or7F\+ A?"5-nGP\LmEO z䗫xR>A6\|^;`!G>$o/RP ț+1{XbN0",L > {,U5$l2:\[}w{M{DSj,>Pv_3:MW%!o2gǏQY/|:s?fxvrAa8ʇZ h {nRp\9L=|Gd 7<~Sh_ QjPk4"X%#Ʋ@&?iC4mg4"L#%;0Zdi8LC B55طvh%/F;ins$/ sd:L ?6Piƴ9Άy4G21)_bAz49Vz W8 -ϑfm6KGOK{σor>GO)VQ Ri;е Q=OǻRף=exIv-Nt "W押 c0t9~vZ x}7#Oszß·_ O*9SRX1~K+:D*{(xu{P3 @aSakjո9t-0 `~"}aL#7U奵Z$^L;] t ש!8_1`sLoAUjkK=`-|eKjNUnp Kt c7F>DN>jO`USB{4I#9p Dju[?x84RA6boH?) zDޔolh鯧Qau,!Q` )ZH %b'~ۖuydn?Tњ gߒS8r԰:icNG;&!8qz~t *֖©k罢HU;}9l!h7;k9 Ջb/-[o2s ^::N>m-809y+&T;.bAsl{ κZ5ݍCX94 3-fX2¡·$XsVJV vskkps@w}J:&PraLDJJGBL.}u 7ho.>ww%Mg_Y6[~l̿=(fZ+>1:b~O_83pӖ'3y~%_ߟ ʋtN]^^r{\=3ydbFJ{g%|(GB n<{՝aiy<uZKg?is_2|օ׼- t&8Bb0m,4EouY$/g,)e\))feP[\:deݺCY`t/ 0sLAÂ9d]ٻ6n$W6;_\竺SV\q `dV5f(iHqD dD%epxn4 W;￾ J 2 - vbs ߤ\q\qWsVoM3o03HWܺLe6<w\jB#sKUR '3?tO* ] C,&h*mY,?ʠ'ĵ|1fLj+_S>vPquin9[?JPYrKy-6kyccު;BVYq$-I]ؗ<,Ns}@SR& :Wv彛AڛY2q@`}k9ߙR{О¡A3zćШĪr!q8!q8!WOȫ!4rr;%LMA)f@v$G߸Z*,D:sjP !p:WR!(¡dF*-mj%mߐ)j4Www:9}sQ'MЃՅ:m]h$HŠVa@ #,Z^LUڮvI~L)@ HgS"oo>W(oOݒ## _QA-d o<,a꘿7T GȌ0s Ӡ}g iulQKLe28`XO_+Zv755?Z@GQ{9n(6{MaH:@VX C p4k gd =3ERb ``_% vb/O׶p& 7SaxQ}[v-"24ݚUy >7Q['*~fkԎD&9#@,`{ш"&M3u-/0'OxD(DejQ71~(z0qyžM,^_\~2$\ }֨jɇFjStԒ19Ǖ%%̩ 53\cQ汥&R:pBdžDGq+ i''1%aY ; ޛ\Hϵ Rgd‚0t(408H:S^00YVh5XHIX:0hZkۡk=^(jF#h&ׁlVP5ԕ Š ĞFxDXڛ$D?pqK9C0 GMX TT',L0TVS S[9#c%tEzBh2F<&b0GISGw;9&#)b$RȈGdQ>x3]~s ޔ~jF`'Ю2= FKõxwJ* 6dS)j,R7ovl?/̧ƀi f On]ծp,Ds5KӇ?Z?|*Ð {[ew)Sp+ N$ېS>~qy@Ҳ8E *Dw-@ ?¡ad Z)Y+ߖiJM!tNSOYGeXeCȓLj^KCn-2biBQ cW3ƟZNaQƼ fMYl)!*ac6VO1#<ϧδ6Eq4Pu2bi& ?Ny< RBgeD˨J{gtPe'ϛQ$X$a(+a/k3藷ScܻfGo_?@(E: ]{q3YL'kd}?,. @}vS#ne<'p)utlҨ:%w88X]rHsk345HWt!0Ls8sb&ynE9Ϸ4_Akw_/Ds? mwՊ /~xLmb*M|a4D"<' 'PfE~q7]Ers[x>/v; ]ހ b Xg#<'jqC}]=ucDihףhi7}~}NpzҼ:7g 3LH$)CFXvI0ԡ)nb&8#4"n5y}Mfu-KTÜH*i CgkiǑNL [N.c2=Z-$$eRxf\n8W()X֤n _/Fŀj ^;A=qLiRr&SL)n!9 Ռy[u%UM& iRjngmy=yi!dWC\PXV?V&-r-#ׅws[_ѱr>]96dnTP{0K/"uAЫJfIl u u&tUjO[Uhkõ.PZg\E^v 0&5ޑم}?+,5ՆMb>\! MQnaT{u3϶j-ݭ%X+#%?j|7mQG(y;قߒmNOfzOR`ROm7ղry1s{s q[1oq`-wK3 ,-&ɲ̲rW?׏Y[o _}/XLf:ʰto_jY+T2AW ̿TA9Gf>ꭒ>s.+{O/ݤ4p gtFjBЂPsnDr$z:b㘀j}mA˝!9By ?_ hq|&rON9R:$ï t<&:h,ݸbib[?L^/P4Wp$t j+;p'Hg \ DK>1{$87InKټ-/+ >! aN$Zh^b/(JVm>elErX_V^nք*#KȃxiUY〚ЮaM,pZ /UΨn\G񺘏ʶxյQRY'!דe@>:&">FIW_N_#Ga ݕo?_QWS[v0 Qqs5Ƽ ѽ=ѾKXLu^#RC+~F؀' x^vHޛqxpuVy3Gr7턬ˆaD`k_őcp䔓k(ԌqLCjU@|)-o=mFWK1wKOŒib4*)qut#9S;UؾgN4}%L8hy謩uH ]H? ꪯupvul/SFΛwgD?8@Q-%;l-ڊOUZc9{@w}|)k;ל]NrjiI!A:#!DE& ƇKmDt>FRؔ\]USfJ. ݝ((4xݚ{iX4ǬM=MWU뭆Z7o5Z)qj]u 3(EZ%QJBrDrf:#Rpj3%`!Lt$JZW׳uia Tna\P9$CZ斤6uwvXIL'O"U7f\gvHA/yGeOnb3ۃ,8t.%n(Ӹ,ģA%TmOɾ46(6#&9^QAyĘh H@QtWۇuC}SM%3ÏjXR#Y[?*?ɝ>v>NP[?)U ^cZXckSI z\ 9#s^K K Mf]Kc@$"4nR@s1LQ@ҰL <YK-g-,5q >:0R7>1kÀ cN #\D8e͓cQzt 8<+ጒfA~XSX\/Jl:tra^dm<~\-{VPYJ*N&K$v7!UÒ` Y-,hX򘵯b$d0HuᝇV]{+h; H мR7\`!;"eB;Ͻc1Vq s1 ы|XoLVH `O:=i6Bgߏ xwZD s4GuQ z{ 9yOKAI ~Z Z>D׶˟;%'svO; #7_5we0.狻ljw[:k̕]͊T?7n SHhZrqz~WWˎw@˯qFԥp[\_ oLzf|Jzl;iQ ;2|gD}Z򗠠8WAQfX_Mha~9Nm^w^ @oJs0T2ut6Jw9JbƎHʸ/lGt[Ja@KO+"8UX M[Fes*Mqc“BLin3ovҽYMʪ~zw ʀ"DiZxzMF2(x|[|fo>/tvO sGLr|<}RdwXkvr_~xu @E,JT$:*?TBӭyՐ}|zrP98è˯MjFwår*nl('v#쏎fi)gOT+#3ǂ,zNSَwZ<2j},>jJaho(; ]( -hg:7N]\nczRʉ:Fw;w3^`;.I Mt^+) ̢w3%>i6i^4+${$o;*I| MH`t(I,Qfl|GP*apm;R֮ޡeT#ßMO8-\o~<LЃ3-k(SӨ"I5&QJ$>Drgiż7Cl }.\|~ s[v~iQ7]ҝ]5gB|?l8Y$x%u,;߁VUQ(*uP6E)TeL.ڹ7SBPu'棱PSZ(uYϫ#Îtc{V5z}- J/h&0ReZ8-1 $s3ltn撖8ԍDv㾪j5;U9t?1L(zVfjwfSSk:}<:|_!۫('}i ^N{ј s2Fcչ-$Fo HgL7ZGch~0rWXf 3ǖ+y?_-DgcRivD Vg:o4GoI++mg&qۙkTdB2 oy4h`^4 0/ 0WO@~3 \S8qZI. %Ab Zh-˅БE;{/m2- V?&Nj_v~ hTyV1oC@ݼc˻'C*:]/ /MC񋦡E.8LriP D%ʙARxA*D*Q:|yo!R`e03&!쀹|Py3xֲ_9NJ%!6 9}1cK(xW9|Q`δ%|вއPNOowѻ|GH .b,b+&5DD mET^Utکo9[$J ;a0cڰ(ДhB I:5=o\t^my*KY۷Y%H 掬t7ϣK~Inv\H'<% T3#ɸfhR#x˨5[)s9b2]<iН$#L$~(RH,Z@Z $V{o~FqoL L MԎaś90D0I~n#p4wq>)5]!|T$LP.DO2ݩh~!Ph#-)`$ N:G{`I& J'"k(p48g)E|J\bT&Z<'|;qO'w V(T 6?7@~?>ث8@+1 7w|LaqT7|U LC7Qғ܎QʵpfN/-Q*/:Z:FϑY^*Z>>MReΈҢ3aLdFrJtu b 9x}Y[1TJQ3$_Oγ{: g#ccJ['g' Mvaʺ }ݸJM?ИK8\KnXv"(J^px @,|>t;}\]7Wi֬Λջב7 ^Q0LߜOIu7ɚ?{GnAN$,S]xS ^%Hf783ZstO_fؕ5=Wb,~MT}\P=p`sUgkݶYОU*UkThHUtdݹ5列vhU]ؠ2xVP} rP]`K'p};0)T+]#Vzʺ{&4# n Yi!)iF-q|zJ59KgH b+b$OvK?2Jtrp[`Zpgkoߜ|M+ؕ6~UjpOa̋; il>i{ۧTJk5i7:7δǼB*[AКq/t[.o_ZǸ{Ͳ[_a Fv*SpI-2¦QNah*ijUCMdžcwm@CGdÙʡDуUYv [RH;3ҳ,24?W^Cͪ )O>Tl2PAÛRVE5zqFNW-ƺ w崦?=׋1X7mufMyUØwY.]=.MO! pG'W&jVJqH*|\rniOL?_=*ٿ${x:53 ߄ba7 {ݩc=QpzwFńToLOyy!UgK- 84}bB pU΃VZ žL1$ j:f RHp> C(a*!R6Rfsƌ1& 21oQ[wty0%TK9A01DnE`6Uڅ%rDQJفHS^3dnENttU]gدo˛r }]:m/='zUWWz>K+Kn$I7\>iIT#Eǧaq-TA2R6iRPmxQ64m^6vQ sz»Aoueп5tD7NvmӈQ'd |㤋t+*~kuJjԁR|8ƉMD/ӒVʩ dH>Y _>؂we8t2V!5v`/5^khZD78_[BYd/ei<[;F4DʨO??#]aq\?/Eo|Z%6b,x\ؗw6e)~h}޸{S ;+*ʵj5iu~]U-uKM.`f:a;jbd0i5z6@4G2fZ7W(iH*WGt oc @5er& AGU$WɘOe:vq I|JݓEIg܍)aɏH朓?}w&-MG!+%VBU͠w߾~>_]q*3!6s(g;7 狀dce ӷ>K<&u8i'|N A6 SK%1ڴ$F;mH_LR O R{~{JX S/lϐ#xWc]ir64`Ґ=1 EȑMP4YRT b'^1x] >#arMr>fsմi\10&S΢ RQu2DdQ:LDTj5Zb#EˁuVadB91Lf~pZpz㷻Xr?>|/i;]Fխytr aNM۴7SO3g?=׋1P>cZaak5[;҃>lŒ =-GE .;Ӳ¬)-PA7K|bPNbHk@O`jiZzq`Nc&1&ϦK`%RBQBS*}cDRI=S~}]!JѣKb %\2qU ))Œ2ÕJ/VxeQX夠*U͎ 5c(o\$nd蹅֊Ve-uјcIA)k%Wi=j3ZGԾ{w#e"7ób0WVcRi~s}f_!D%͆f&/'t>ٙQ\.|qǰ.WhGX0ϛ/V4k=RP۔و$.gέB9R0wMN[=qRkL']anEPӸ`?}h+hYw7d6{oo1<7榑YwE 4i/֤R@zkdžld6jQ7|2lr+u.:ۏݿn?f׬U#3a\ <-"!)r0ura/i9>uV@s{\$./2Bg]"ROpe|^"Z$d)6}Xc&"S<3ʇ5cL)Nܰ :)/Ů  zd /볼fds2qjcR*|2~f3ǵ%X+ Fomku5N'4d84ӱ7bsnD׵? o6]84_dž(quεRFMՅ3&6'qC5g[|t"#dd>=ڋf`Av6Ч 9O7S2iFBM2*8JO4kvʨh5}9# 5OdB1w[ۡmͼW뀿G>I:R|TV_5vՆ/?;_DłRKŇ57KY\\nTv}^*_5j˄:ߖU]1w#{6x^gRH̚gW&3"2^۠|`L4^߉z;~J5(.4w 4ʪ0R \.PӚ0B}z51نĝM\2)HA/"GڠBQ3+fų $z @1`ݣuSٱ7bV9|=du!U)HJ#7onV~^6K+O[>:>67#:IKZk.y͜Jd Zp>F2!DXf\HqF *^\[wwl WoJi3^-"|cvW1(R+{ʍF$L|ft 'a~b\OÚѯҢ)=^Ag _9x1g p9+;3>әڴln]V>7RIeDgGߣkXoI~՛%i5wC,B h0QQe2afi9s*̲* ~Ng>>^EI_8ԯ%` r0q$7|RoBmUgperA}KS 0l䓆9xG&Or?ϱI1fڃ|wvGJίa+_U&\N inshZ@t"ГM)ߠhOI.d9ٍNR)?Xz:gyᬳPLkG$"Y!Elt"dLWb{/AIի:ց7x82M 5Q>&==~B^qd567V+G*;v6-VEk D[a<z䠙Q5}=L'=nw{|N+iw% ޵`I)c`!xǤ)JҎ%ĸIH 5=c߉zu{;:] Jܑe"&9O)BwjNbRïMNU\p?KNMFKsz5#=M/AhXo~^tZ9.vۡ!?T؇H]$ڌ(J 'hFWPOr}m;bXKׯOq=g ue躇bSB|׷x/7>{+=ևE[GbqOONn1a0f.]^.w<:9;dzQ*DˢI=)= Ef=ږsϱDLkO+Y|`1|ߥO!}я*h%㨜u0*8G˷h>&S䣡-YؗfC}.~H8DtIrOR*bEV)RY[ Af^~ɰgGO}iM_/YˍgniݮJV_ỿߏ?-?6g#Ћ㣗a+婓j#Ś~un>nh|P+FhXcKJOD`1 1Ť}5|u9˕ȘAГ}eB Yf@KhpAEfn߯^[NGJ֦D@77?}S֒^qJU\n@,`f"_"D̫L-mᷧe5U&]+ey >|J?rYƘ j WA4$Z8/C9)qr.1YK\dQ.i2 iC׹?ptYI21NiI ɽd2ΖW}M X~6 Ӧù @v,7},""fVb$M$9t>D/& vG.)%WSE"gI% ɕH.U;L U6 HT#+g!0&u/$͠`0|+"sE<'C/TEF9q&  /b?]g@o=غA2B1ݮj7HmmBZ}myǔ?6kXCn\rg  yf: ~s2IL{ݶ{1a(&ҕ$ ڹW<)֔KZɑBwq.@ISxN ?ԙBVM! WX;V  8DpPpəB\V'Mk3'on1qbT"1:AL+{1eS j}5pZhF$O)B"pF)#ň28,&$Mx #B283+*$*Is>)UyIQ!,$ S$aA[4 nT= TH4G''[0ng̚2\0^Xps 5+ͫV Au͡*8/ VˈtX{ 2rcv_al~By'2L2mP)g>**&B*]2T9#()gR +G،5EH4/sZ }۠&IT̿y)es X^qƴ@ӪdS3\(DiΫSPkд&H6>*/mfL;;CV=f^v|6y7_dF1D,H!s&M4R!T !S^J8"2j$oV?ߛQfQnOxY;TsRf`^gMDTUV ꦍ"*J.%eQJ BѪK#Ӑ ?ざusU#0} B(ł͵ASv&4V%AfiPRcE=N0Za2V ƔAk4KNxcr@6z1D\'N Z7]ƽ2{Vl6Z\hלKtQu{v^cvCHyҼhr~x28)$Ml;uEw]6_Pġ7pF~٧ޙ'&Q+yj`:wlm헺X*|xDqqnRHƑ?pg}:Ω"t 5%0YUm|b owb_Ơ>?,4< 3*i&TY1VYAU f't?zs` a_*BeAdQmi+rAF5L`8[@jM̾h0J֪iU0ATizW'޴b׼io ϽiD03.M~9'TP&s>\^foV޲˼e'XNB7ޥSP2*0~r 2d=V^ޣ;]9%I@^y9i͞$epk}H˲s=iYVmӲbJr`Tq68r:9gʆ'",I_taI;{/y翗׿ONG c`‹liȖ#/(!rm:^!\N5I% DGFǜ>E,H 6۰uӒhNX]qIR`PH&$2z̙ΩbLnSxak% #jՇާ>R@W[%(p"4@@Q< iyG*;ôh3".EJDpy Ds(+)م;H%ew>fw4;q&彇 T3L{$"lt{k<5mVPneP ["bT;ReeFuX/>eLw /~fw; t'6A>ُ>ĸ\Cj^P'6ٽBtFl;G7C!{,K!`ZWcF Xv~G.c;QQ`* ]hped7;W7||ԹՇaƸ~i+#ݿEqx}ʹ }>yi9yy%@) f|F 5UKlUu%G%U.fl9ob7%U.#4NQ&儂7(39NBe >]vs%i9O2`C|n[cM9)X$]rSܣOz&cх6,>-%EK*h˹ @I1vQQTUU/X.^֍c8H!wI)D[qClN2JG]N~)7l2'6F#>?_ݯdsdo~wzGFyT߾MɲD 8)l(>oV!)G:JiTmPFJnL))Ew~~4?aOr zxFD ]-߫V8-`&9|}کoڥ'Si#KGӮ&ˢfFʐ}EIo9E&\IC{#Xk!f '6j.3 YoL,QګŠƓs$I޾+o^|}^fP^U! #@ȧY !2(5AWq_J#0'u>bEcB"9WDq[:\ĪM&α"- aMB{6a?/WsqY\q'!݉h`o3%}=+{8w0 ???ՔXp 7w|jਔP9\|d^aA$ ,u(QGʢk#櫂ez-'3j٢7ϛm}/*`(C*3)*W0]몥,6TAAC 1FV\8AxIErކZQ1A+J;r lt9g`X1"brx>BDՆaҕ9/$RP?qo][?"Fe򨱢z'!9WlǐPV?8!Xbo>.2~V8I R Dy\!YidW8¸8CZ/DK%#_!:b~w)x )-aUmgd)wn-j.*vkiЇ7yov4Qr̠Ժ#Ce\Eٕ~.pj KxXC^ՅYVqGBUTdUBE߻AF/wa q"Ix|~23c$"e=UY Bk?5T;!"NޛZ]Td)߉H$y$.qcw?]WxɄ9 LhN=#fJjȀJk y+1=gmn,ɍq=J&~-:4cX8Ye%\Q `^ŏ͋PvDjM䬖iV<6?moj^cFdDcXYYE5eXf=x "XQ#R/ˏ VG2R xDwm_u0@>59E^-K&e_KN %YɒLyw]kyhf8/ t$5[?8tqfoPꑾ!Y+ 3/Wh$0ePgh wresb)9+ldJ^K^4n(@%C8+V[džfٽJSJ'r=Bw"[|WL^yLwmP=1@ckH? hQ|t0ڝH'e[W= U;io<-ktDG9 d%kF$:j$OX.Kl槆y>}IOXWko9}8oK ^HkzhZEZC%@sgCn<mC7vѷZ+ -G׭%U"B:WJ^i!JN&W-a_eh]$3w=fv'K:{&rŏ,jϿzl²t#ﱿ0P/;F}Ѱ%Ĵ,@-긷LLPJ.3<A93ԝx*pa06`7$6ÃwTЦ5wB>>z׋Jq/QgJ)K=7ӆJC[._nt-{QmN5ܲqB֤3hFƑlġhU6bQ]š2Pqn  3|t} S;ۙōGANX %9^ʍ>;Y ٲFO٣49P=PE_*BԵhoPq (v>r4cU?ԦoZKvq!)yCYKr6|a{)@~*Wkn쫫U%d ~ԉG'IMVA5`mYw?V=F YgZNۚKMUۧH'kM 8ED'CJƒJMb;ZRUzxNϐ߃@^zz.`F C1Ғ'j0zi}cQ;"HRIbCMo¡j䕱PCUҚśE}bV3ɲ8c2ď,~^u`өٙ$(Vye4妫?TSٖiH[BC}o. >n)"ܰUOB`ɱ [xB8֬4U3ݤv%ug&=PRy^ƠT5mG! aqahlqr7L 䁔h{H#-![ xIB!6܋/Jp7kW'!;hkLPvT£ZGsMUs$an Jqs$6hց6SP콽.y].T.<ͦꨥ6OgAҕ 'Ah*ZG=-ѬgaQ[WO+=ʎvzVL*jېw^"ͫ'+#>~wOMp.JAⶔo*@T̻d\(U?B)}KjQЗ׻ۯok23UZ4~>Dᯅ Y(6wɂy=61PƁG]uP'DnYeL_ FcN8DPցr{>̡Mze;^LʋnEvTwף_fvv7 "F|p30_ kfBeO,] ߛn*],T`" !1=wNT`:s~K&lW*r-*Jbou۳gHHHO{/,Ҝt 볺yIismU䔗%%By!4  Z`s,Bi>y( nHd@X R g)$#\RZa]9|h'p׳ .Ʈpi#q%95Jy{AL%}tI8]7N􍫗UzX-A]|a^84~ayE-Ϸue.l5 hP=[ΙhpotBn"mwԠPk*esq4:4SCs"hkq*q*W < J_Dϔc ݁kC(809O}"7U"[C]x||.@kn;T)DjneH{h/| ~}>pICw9ޓS,8_nru)]I~Dw Msf'u5:ʹ9':= ]&me*:qUkR`jm]L?v߿}#DQ "oR2+#̇D`EK^uAyB~u1 oH 2ӼmRq`(*ߵ\Cmz-ѩ4 $;=BcXs5*u.pH0 T'MbZ1m=,J47͒@(5^ + "&ֲHuVp|) ZiAz&֕J i"!N1JA\(jB.#JW8JRЬnuV9Ycges.4rYO&;:sD_!,.l|oԧ/o/2[/x+VjMB:& xU(HİHu] &0bA仦IF ˨dQWR ĶƈTGk1*j   eE_IJ@,SGjIm-e5z! (V &p_2e9rM9ե0;P8?ʍg^0WTc£QIKA/~Y@ћ"h9$2Gi/įmc[4hj!*nT%%* 0h( FV3<5˴ĽlHM0`P='B}p}7*'?`8+銃qPoC5Ԗ{F]oY{4(hbuDlIO.TĮWHՠVv6΍[V8*K(ҷ/$0,DFhdQkCWK㞲D[ QʝhIP-G]R'7k|0nf#)ۏ(GZ>=* |p/?xox/5w*L6? wysSajRo~{|;-%n7g)-y=3v: g8|7G-?=z7J4;Is\7Ԥs2}MΨdKO% FCބe5ķOsCW[vd TW&B3{zi .X\qfRIQ|'F02̥$D떧D \*VrAc,Ⱦ JM҆'BU*F_SiHo\s ~W6A ̔h $ӏateX`]0/22A]YW.,eȁvç8ҰX_pk8ɛi6y`(O9@L fǺŀx7  ~ͪjxf '9}?ӱOw/&_ـG}wQ^N6Gk]|Y^2%Avm+;8?tC߽l8\|JvNaYlrM69&4b34ꖖ`G\OFrΧ,.Ƅ5Т6g*ZtmOKH1VQ-r58@IL,p-r[>І~;@X{os/̯ :$XfKtq:q:W;-E^*D: 7y oEtbB{hs[.T%[L-צ,CR"uIʈ!H$e{I8Xf!QF̄}^{7|uQG9WAN*r9%Y 5wm-6=f012_R*HmwNpoI P(NڊZvcinKZ5F`eY!j.b%5z(J,% br AX8)9\] T7=6\A?YkZ6-1㶗,K GG3fG艀sJO\d2p"2"#>nqZ6Vhįͮ-VwS˧@r iLAS5gs?F*0nxQL`D0q[J*+|~??Q*y&c&yW!zVS%Y1`HBlDU1 k0C>c0H$2 =DaAY %ϕPGʿ:iBfᓂws{7goΒ_fjx:O 9l"M[6. ͶY^BF(eq.QY.^qK%dz G1#ߜ1^dsQIk|{lJm\|9Qyq;gocӕo^}./ܫүJZE,Buy&%x,5ƹ24'&#TRce`?]YsƲ+,={#H/򃎬8(vJ.V'po: ^$=_tt^DX /G *,ʙ`Prɾ~hm6A8 xM;;gfWմew`|k1'}`aHMf)W=Υ̡gJp+L4C`c3f 2]#9L8B\#Bc)W=TQᖣ@d!g26k/3A ȿ ՃRLoEfÎƮpr(Jk32va HnbPa\*U ;)-%v"޶L65lJg=|닭? TJ#m6iN)͊Į=@XWBn Q p|ej>RߥN0f\`u̟Ql ̕gΦ{C "ǣ kqhȻ!18xmP[qo)7xJO9[C8U#rK>ٌvEa )9*ƥ94BcI6}>YѠy3y G?ZH^9HUYɗޝl8-!9Vd% R$Hj\l,иN%[ς| Ne"S2Znb퍤38 1N@=QGQ phEJo)cOkjcD0<GR4:". K`NDL<6F;#өL(2ݚ(V=&ثq'\cbއqgP8G1\n1 g S8=]H}`h<>>#; %Li)qsK.G^F,Hz^myr'cK5\ypg$+}m땾[P@cgj6GO(S]sШr`,&al ʂnۥʁiW}:c.O&`  DZ\fx۲)ZݿV@yG}!Ll?>׽/վf< k?C?Ӣ22u]oopY`mg|&䴑ΥՀsr-m*XscQ<+BXe֕3q59C`J`jXlTJB԰ZMkZsK!R3n-%t[br‡YE8ď| ?إ* Or *Hj=Fl45Mę.0NǾK>sdw{hwM7;6B.WRYʱb[L6#{bwar Jw_ C!H\>u[wԂKN^ܽ1W& Ga.R1LpvaQث[Tp(?D*ݍʈDQ]>N"tmg _x>Hdk ὂ?}L"eߗm,$䂧VJա2UN9watHr$K[;Cv' DEef>=S_ޏE0-'z C#ao \IFD0mݴS- KR\"9e K1B!v־L%Z Jm<|ewrpBbfŐM!hI`oZZqg;ABvf]qDT1OR:8NHZ"Y${XTc*>NJKv2WM/ a-ݑʫb XpE ?2QmϹ2Mq$ +)|? If#Q"ɷα(9ˋ: BTRTfSKC ovwDF=j7:CP_zU CewC$Vf](sdVIW'ݧX  ;&73M 6`?:{rb;&.Lt}7X=h#xY̅^u*|vyoi|lE%DLhq[k ŵZ2 hEQ+[QЊd˃Dψ~K-2s9QH f9ɝQxWw>}Մ̧phsWͻl\~mƛ~x)znxo~zys/~^\pȏp鞿OA4`q{C{4=44nem?<_jޓ `~M7)Oq<5#3 OM}sqZ~{6^߽-_D{|v4{z K'!/㜦@rMlE qxz>x_K(Nc8xhw7@"&h\`iOͷѻCCU{oê…nڝ _o_/O_Cûן7#,pa`puzI݋Od[oz`41E\5'LE2{w; ޓ; z><03'v7}>t/Cod27OLxRƲ|8+x0tq2l'Y|j^Vȸ"_'xzˤ]ܑ! DVBa5h ! 蛦Yg(B1 ëҐ 94"Se/DȜ+>u{D[m Nbx$繂 p-NO(NNL_`MZGtF$|MO>\|:6}؛6C/x(}x/z|1z|FVQpn %[Fg-PE PڵQ>~:ὠVph)/ < }Q/ /+ܓ"Or e4x#F%uDHAeTkʐzk/ o3el,P˞ƚ=,@2!pAuWfbY K ) 3`2"ZȳFLGpPW3y"}iSX ]Ja@`0DFL ǣsuX.`j[-`#REX-3qk;ft0rٻ8@9zMJz׵;0qLPS3>]ŸYҰv&^ԁC 00i2!KKS*f\X_K22"]bխlmܒԮyN@)Ku~檼^E .fx]}*ͺ([F(UПyzڬ$˪Yh仰2+|51aUvb|C}:"] >2:*u R"m])-L-;/gjxr9>L.[ +*v#ƵSi2}PfyGlEnZkEr+=W +\)Lޟ M󢆻ɰ%|6h;j,^#Vޟ.UvduJZ.dJ#|wYG ^Xޘ P5Ϸk9uKT:W:Y|{Ѡk-t=!oBf. Q"DI=DG`ɛodEh@:jBaIv]}?S h^13(z[`E+/_vEe7ǽ$-&FZM\xԊUǧN,:`u>5^c'/v XI?thYˡ"ײlugj26sJM)3hzQ HZ5hB^8)ʗUV6(HtYIMqs,?$/Rx7kCJ\µA+i]=,\KQnMZW>%W2ĉ K\eq,yE և-7\tj\"fIyaW'3qUWk;LLO2`N.s(|[MFLZ'Wr$-3r UOJ,uZ aaua XH,_fm۲9WoyުC(g;̰*ٶF@23eR3 Qn*425S)w 7P#ҲNZNi+Vλ4TB(UUGfP\Z"^nR{#5m0?~zIdKg[ Ħ#V -RHA?"/Dუ좥 k_l]e +VJRiRh3&GR1dHT*&U Tp)] @j. JE=`RuW6&)*i> QO#9OTK?hڨb/U,eTZ}t,Lփg󾛮~j|ǎ\vNS[:(׌_zƿo|law4-yr'o_46OO>>Ch}\&nmp 9m\i#]X1//!fb&H]mB:%d'Mw(ɎرlIs~H%"g!Cr8t%u(\@۞DQRa6X`}0ޯ*ʲ}CrZ$3f#=۵6=WL8̴i#KaZ5q%.6*ʌkTҜKOoI-}]- B%g7sW91ZmiEmBy6L9\L0cz>'BHؖ비HdTWlVbRɃOOIERA)6!U% 9P6u@ b(AT2ly@fE11=YK9y,’@ B--r$KQdڶiSKKl+%.! 3ՏՄ)" C}[B:Jx&3m̩ۧQl@Zj.7keg6ؖvcؖ:S^o.ZI4]zzlڻןK[f_W["sE:ώޟ9:(^5N΃7oF#"TF};E&Ft>{mS^Q$ꪜC^燘{U$^&^U?մH77訇հe'Ipa 3(̍j'lfzU5(8D!k @;1TDc,̀pφFeE%6;*+*0D33R @w9{(R~Rl&;R7I *QPmGY@|4 z)Z~ZzNcrl9sA.FنinjɅ [SnK:SzxGknR8"Gun1EuY\Wi6r('≲*d[<&{a cFwb}D֜(?9{d :%ƕHTnշKhдf+qԃO a Wo{Xc >C0ΙK Hq? )=ިҨ bubВACRgby`Iև B%Eתq-03\{dc敯׉ɊKJB3 gt}MD5+ڥNsxPI,m7Ѵl #W£) )C~9Ft!m^]Ll!H }k8~\ݖ4^5ąw7 ̻ 4JˊȾM[x~ņ=X0Jdg" ñc't;꺣8ý*"]ià˦pn1fk絎rL?[&NwiWМ/S wkc}Z (~j8vk J 54c9*=Oq}xp}^C*ێR:9>>ഝF^}< ^;-cÓ{{8z)?}tOgQ+G`xqhu ܜ5(+e28:я'ɸGdwf ,S@BIBqm2TY?>I-GkE6Fq B7E' Hp | `p}>PKT|P~q)ETT7ZZf0ot`6?%tn1=U^Kɷ8oM0??c es ʐܦNⷮ& ۪c6 132+ \"S708Ih҂Av+ ֹ)8նi7۔`}{d=~G 4%e9N9W0u;G')lAF", Z^αd]{g/C3!a}Bː,l65\_shk:v0Dۮ#q7$P`TmDS>]9R7%YӁS~knQjϗ CXi|/ b/c%dq;px]\#2cU1`n3 Y;Cs9Ŭ WO)sY2w:p(5 [3a7L,M2/I2GJ" ~A3en"ޔO؋,a7X  59X Bf)nR_ee$oQKԼ=gӚ%?AFE4qv6Hpb5#ϒIw6h`,is(X]c,eX̲ ra fy(nKJ12U `&*eB/|S 9u<[UnDY<,bmY/ɟAPX?paEnI'D7F3?ZXX3 3, ^+btbg0PQxPDm9ڼED#a!v36S j6¾3@1b97{]lF1Oes =e ZQ<>k>-/B 7k Q˰ E 4,εwzɫև{N:yU'0ݪy;0 * ž%<Ņ0\ȄH95L! >VaO3_:8y}%2\f12l*|v푧/S =dW/J5~YP~濞^:Zv??\ݖyuHJ\xw>:[8T?Z_/u BO3W0N}uٷC1A1h$LC#p1Z&Q28)Nۭ!LN}mWI:$΢n~z=2f!^GC_DwMwۧLN wG5߿./e?NN./N۽^}< ^;-cÓ{{8z)?}6IwtĒ ;M΢=;k~i'RCJg(Sn5 ff{'a;G;m@N!I,q4R.e6p9S&ZM<|6I{JkB5e!s+m}7'45R2BRLSK$orV+ӫOr=@V)v`؊tb3PN5.eR&<khF2pdrƒ#H;^>DY 'iuŸU&C G^t3)(z/̴q4dDn*MR;|dq?iCu׷{ʞO/.>O0ئf:ht|m5wysBߏ{LѠ}Py| l2̞ǯ/'6&X07L gPF]l<-I+0aay*̌MQZ#&C#r=Зqfy?*3?A +H|3ՏDX{Naq.\+eӂ9qe*lҸdشl ض)NaL.Šnp|K:g\a ԂRb7)*)fq{>"d6,F#WDom)<$gtjQjLƖS1/}ͼ7t%ܘ] CyZs^xLEwRkU" "_PY;3BƂ6&jSHɃ)(nI^aiOnXVAK)#mZoؤ Z2$B~m=7M8{۶Kr2xV@Rm\$8w̿ޅ4oFWAxgTڭ.'* 93j-R 1Lq0OR^lJo9&tuBrQN56V^/FXާ=9oPLAaɳ%Ο/ LpժIWvk㊟z^kzXo+qzh>T7E_|;|hV̅q*bT/8絃׻~Ғ[Òr=,pK6ÈagYd}*9hXxz9qaIQqꓙ-^6)hI`\flL 嫕7g0O|p1b{Q%H ^?<ǔ (wmq,c$TWuE,,vt U E$ O./ý%$ElOWu鮮ZN"Szoz5)*'Wۑ C&]oP>?}svQGi$+19c_O?K,7";gȗԨB;ɠ {S[y-;d  %n%oo#W4nj5wJCB'haU[u\8 JLbWH՚bvQ[C*g[\A lZy Uzˆ8%oEOү[b?/t~737<˫'.rï!R핹~-}sdԚ|:geG[OtW&ztv~qɄwh qJT`Pe봧q8Ow|( i9d%(%|r"c[60OCZCg4z,L~ yA8֚lN *t\$Sra ;=nAӢ>ystK0ȕ"Q:K@")#WŒ\9 rSFaÖElI?#TE DQtFWk8yrZb0(I#e$6ӏm%[*ԅ6.X#uh8 0Yuyt3}\} ).uD1e5L.lc}W6rtNu`EW$*!a>dX2Pɠx/Tg`a 5(cX>T-mG7vaA2݁h#&y1dۼB N]\;S2\zP=>0wQw(8+|D)V.6ЂPD)'LND>׻c&6ޘ|."zK>:@>G8J>%P>WfUbAU$}] 7oQ[U|f+%M#k) @YE9 r~1itFܤʵ(X^gGN35kT{ /1}8`Q؇ܶ˶N-!-/i@~pLx^M 18kC#B!:4zX-\zcA0u+tPYG%EfZuZSZyBpꚱYVoƿXdz"#I􇀂v˥]"I8&s:k[َ8MUkSYt3}oc%ץOkk/($Wny~7eNXs?/c831M iݏߗI8??pVEVXz|VŹc`4߄Bٯ}o>S $;񴠉NFG'!"lOQI*~k16hDclu N*?w: $ű+;1wVt4yr=/•e-_}ɞDsA|RuoDOg [xo8B?8qwݽVؿav8 vs~'x>1+i`Ù:<1^˽c)ul"lO3ЁM7xs؏Yiu, gwD칹2Wops6Åak͠iO ՛n15,.KCFXei÷[k6wR4eO)rͳ\:%E!kޗ0 FIh^&VT QѠ)@@60 Tw eցO+o \d^[HZ2j9gثcrIr"%\|ߐ1XO| A/Q֔`WVw$/;2{[AZ!TU"Dh[u/`̌bCsȪ04CΓ[ѱXrу.& R I3DU_%`W[joT>7 9`ŃPoA()3ή8 ջbPޣPWgw34 vfȕj%J<:/VͮwT8ELܘis.551f9[/aro-$aAV;_j K!8w&lm\1fo|!8V<Ѐ6SV.%r2mվR2]|}kWSHd\!!WHd4'vV U5q"0e-!䋫u^} Zqe Pۋ*-$7jߏږDK50kk ZpI`!nO:bˊ-U&V] k!ΘUAj!8Ȏ%1+֊W  );1RkyosU?[O:ij ބbSěWr; .NH6s4[Kdž}tAB,Ȅfrb*׌1JH6ȍ-Her/*CtJ(aeEq)ȩR@0eS,3B Ti/+: *kJ$w`$r-r&&8² AI2;#_J(nK_ f r -/o!*Jj_]=7|`8ׂzjMRvJR+;C+;C 3$N.n"^Ӻ KEpʚ9!&:@ZoStVeE`};Trz 4O[pJLW~3^dP%.6 Uzk=\VIh^'.$% m`y~5K &ӝ`(?(W΋zPb۵%,sx}/y}_Qɪ^ΎN%.ϵΫhRqXk,XR(D96k Vj7!Mўg֝6m^}:]ʷ ÍѫS7Ѡ'gY6}n;srSFwGrJ&k(Vc3/DOe,JওWTj0olJ|<v> г ݉i6h? SePX LI eOAt_5g'2F8H2|HPslni%Te%iuرi=k>ӂa2.ƈm s֛oΪwifffaHS2>luMnݽܟ$|pά/}Ӎ:Դ⠰DTz78.Tm̯J(#J qVͩBU)g|F,ZkT o:cvivDfSS6uhf\U& e*ًe )v!uDq$rL(1r!D䲭5J @*بȜȎ9H`lP)sNo{UivaeF6j"\-̙,)PtΔ@!K!{̠\"w-~[)Y]O{k l1n*u뮆be:K.;zZ./(Ekb[BRF#1k 5Ku5  Ĝo$pP \  M7\5~ɋfpZ)ZArbaYK?8>q{k70BR/dRM2;R>DR p\6z=Pz!hcԶZ{tjƣٗLj{ kugmWntO7~iO>[1J 8J?w+qu{݃.j2 rhk :tYCP5VoV!(?]壳aΆ|ʀ:[<}^aC>eLC 2pj2N7RmM_~> xX)} hf¬A^O!L/otc[RrzQdKбF=WJ^0cJoЫuD:{ r٣QMX9 쮋㚑>e_JN]]pHJV@nqOlLݙ/-qy-,Po<>•/FîeB9!KV8*./Rk@0vqeyvz[}_Z[/ Kl Ue b8M7 a5!3sHg%?L)/"{-)˼ nef9@?{WHv6ÀT Us̢ak[<\bS,P Bˢ`"H[\S>fnIJ&{?x3?EVCveD4"rҙQ.Rr1CO-H:$ae)(n>$#DZ2}BsuakujH0E_>LZ TAovBi%1޴) ~e%neXm1)hfx$̎W[^fL`NQfQ@cG=}Pz ,M9ܻ(.^0`:`!nT0al#,&4SHkI#n_AN*($놵.- cjÂIv,x0!u-9j;Ds ,AŒgS%BmɛNk*,YGY7qȜ4̻Oki4|3qS5<ϟnhF} Bk-(hW? '"y@%v@ qMTE.S@_/S]L+:V cxkݨcSiTcG1`͸ti^㕫}Dc$xwNP{)S=bM4=u6fǣ]qUҒ,ۗ|dIoHI[6Rl4K/{0V5-@ǚSoX:]>U𪆆F7 VQ]?VZ=Ղv|+9xQs׶_+hqM5;5EmT5_*z6yஆ@h%a&9-њJ.ʬmz.{zD{ 3]^fZCA)5\jV ha0Dc^+ 8'n_2]Ε).m]ܥ5sDn{ł,<Z PΨ備3,,W|֤^H<]jxi:$5AR\$1'%DƇp x D!GFrsM lvqiw$<}2mQ: M+sZ r8:sLlvBgQet6kH7 G{ZA,>S{HU}ޣ}=d@gpoRIܷiQȭl{%A5i[kN5"']k%QU$  uƦv6Oa[@-J N{ߢZ@N%}/v#I9c8kBRZS]Ƚs_fT!ϡfЀh.tYU4jmWFDӇ!7_ ?my5-_ JJ U.qԕR٦kytp"kj$+mU򣳓唌/ )76h3f̴5< rc(a(fG93Kw#Rٛ  aBN$hU s6$D]0,I 9_"O(y.P34S%I%{ )/QQR-&I;)Q$47utTM) Д-9n)ՠUG} Py<ῳ]Tlߖoj8o{j[sAo =sJK{;ZsT0o>) baߦ7moQ{(.v5o;h 3Q휥/Y曁mǬ4b˻kjvdUd>ƫO׿7+&N_c 1:ds|!"D&H6&9~^ׄk KECn41#;ri&n( ߰4GBJ J.9 ,*8$Ɖ<σup%y%Iy<+Z/ up8֑Wc~I*Lnp4yxcR6I%c439a`pc (iIFYNb#!-TбkHnZs46Vn ~ ;Jg3dVs-u(9 N$`[M8цa"\rcQpsU7(UFeF/2x}>+>uAXlYzs (>˷D!"Bȇ/_ ,?COwW>.a8͗J< wW)Ί?n?lx8{HdӧGW D\=xƘDn. c֯qS)|$tD)O-Ρ!l50$fh!1gTڍ@p"G@D9]UFk,?;đgP,% liEbJVR^W8(bspl{[A$` Wov8v ~6rv;h`e#)X9dTP!±hGN 0ejDaD.|'{k^ ϞGfP͖.x.xOL87F/V͚:,Fd2i7m3S퍳>8@d;D5,RM*pnj/M%`޴W*2uXWd't`´v.<-0;a`1M,8 LJ9;}Ym?E\LB\Lo=3u7yp375w6}v%~"&\!A2ϳNa) 9oӍY[r+iÚqE&++'^ЈRe:q"pi &r-B1)!`ԉ=CMbc:vMH5Xx{wz4B]b0@qPHr<5&xō%)qB*(= :ԱD؆Eɝ"MRCt, U =3`ԟS ӔhiJ˦iwUar*RRY6x]lj$%kohŜ֐sc(?Lf,,v}6gb9c6^"%طy0 mxxZڗȮpoYN_ź3De?ɖw^~*{[JT+8i2SY2SRC)- 㩜(khjLN(TeqRISh/h(v5RҡFO5(/94Jmkfu8ឲr)rFZtSCS؝j߻ 睏֐>ihwT;E;:W!'Wو sW!RhܯQx|;krYp h{jn+#XCَ>kI@S61>& f3'_bBQ\g90i&o:=kh2u}ZEET}"'i|iUq$ ի db܀KPV9Ü0\14q x9 UBJoZ r pZɱZhz*Rh!HyFkQ\uBSH H*4%qwi>rd$`TW"'XAJA8}BCdZ ? t>lKڕdˣ-|`;9 !/|vHXjIJ7"ׂO$Galx?~,|չJtJR{7Z*TTɹ{ z) O?:Z%E4FD1wY7J(asOX2WF.]܃O1YW|a˸#eyI [lϾY1VG`=x kx PjmѼ)S)bgCACY׻:S _s kxXw{KIʦ;ۏ"ՐfjE.?5a-S 74TPX˪ <\Oy19=f v Qڤ ( AY]j̯:3TTśtuD qz-\ cRp\Gq]לhk5 LqPu7Ef7ܦ$5QLL6Qݪ"cy%cԍw=LܼJ8]c\ܦwo7bʓL(0#Qu& AOqVl.7bvDB U˃x .^}E1OTzb>sMNbqt{Ŕ`+k鶌lMUf@Ii9<3@~"_ʈ 8Wf_;Eھ! xKÆce3Z0,baP9P5nw>C{QWG42#RP* RZ&C&okihQH&Lf$B$˼Sp)c+CJ+Vj$YJ % yM3YZ"˱&FueD d˵}Q8%I?MhI7ӣ4ۆgC!jp  ,1+4)1[Dw[ƝCIwl3ov ͐6C DHѩq sBP[oo:8"kKýᬭ'6iZRp?9U+žȣpQu+y{٢mgǴD+J@m(f6+RYe{3&NBbŲ4%DT:=fn*A a, a'1LVٔiCGHa Rbb3)͈DilR q0$t$q{AҍueKLI"iUR ?ʂ8#6dDr$x?]^j3݂ 5^6W_y#˗7 !!mYw$UiPMUUi-R*)13Ǥ;@q(o^ki:go_iTj2 ]K|蜡wV!+w3imU5(˅h!X@nC&A !'իbk!yˈI|@Wt9U;!nMWS;P4;=_Y4˻Hݒ8gpײe%xPɜxLk[Qx;*T;D|RWJVٹ#=ämnZZ9tRmk~ #Dp8D6qL](L ݦgkb.rajX IKjX Aۙ,!o }'r!@ES!Go%:/>_>sPyx(ԏOu҂Q b.u(u x_r{ZtHgBe/H׍x!H$wRپ6aϰAfxDq{ۗ7a4x!xbCw8_si*Z]E~`Ս/뇱mxBB`;0"d徚3%#Z \ϱ;!1~N4Dj"QkdDc\Ԓ܄KeCQAfKC}D[N5ŕC(E T嘨!:Z9?/xao8rpGrpL AKp FXFH3Ϩdi\Vp(LCMvSs ^_);4$ t?Uܿ"txSm$(G?axB˲w1Ε_xD9cUa5e+-T\[dUe*Zj0.`|WQO> _5T)knf508޹S(sm5Ni\ 1קkmUʀ-Ҏ|B;e_в1%cKśXƕ[薱mdkyEl-}!X5886q1@]fF8eFEeHCU =߁mŕͳbG#;.T n?t4f2 pT&Cm^8 Jʟ34A-T/*2YVַ叩 Ymt>ZY(b~4-I[-I/.m|^3kpy#y~H[>a0Ո wզ\̙dm(*H[c7)7%%P}k@i)ޤseJ&)s!DR:iCNgRl[[, ߥb䈠bFiOXza,1 T3o3`iFXlsI,VTJ,I&zvCvF]uS2V:,YJxUqMN3&aJD)a)23%F0o)8ygLAeVeQJ2f洧QcJb)+*z151a(9ƩʨBk+PFdFP.Sʐ Xxyp)UUxIPU!IS1aKy @% (C(簔rD >eĒ#y "n/5{.2t>U1yT8 JR!ES+eN#2M3v %5N 0VSo NYJ$H/_ SeV9K{W{~+1,`QW񷦆ژ 2^^5cҹ1ٻt~3"1w~r3f#uNM_L/I:iMw%\zX.+k.q#vT[UT4Qo[V5G`kr 8Q/)۵qts Beז"-!<4vl~3ZOYNJ37N7h'H_\њ̌` | B7};^+K/oyx`GY<+!e/U}GܡmzJy?ꐉ[b!yY-klLkgYlFǷ0wf3% w<}.%!Eg55;!SK.st'PD\k窓[(qsnW!Ѷt:QK")I7~B[gȃ$έtVY͑!gs !XQf=T[PL$tQ+dA<Y \~2szv~{Y ?qHxvwhm59?\o8% LE#3m)XzLVp$M&_Dɧ2 dp7X/Trƹf4ptDbuY|Q§XLS.SbbM&=gJ Vx7xpRq!W⌬w7 So3VE\B8AsqnƸCz:+M`VM{~~\Tt[\ٰBt|NOfwzH_,xE "DK?˻>ut!?`E^8!v7A:~X#c?3Z`pׇH-5On/}->S؛> O+!DMUZ+/IeX:,ֲpʲb Bi0Q]⿮—%ߢ,aaz2ӷ`N>=W !^:+\w(IV1F<|H{S #+B8Rj}xCޗ )Glo߷d>b%CK͞_UWWuYX>Fb+ G6: 5Yy7o߶ʞՎ7'F5:2 ޽U5b Sݹ8d L6^K+F Vg=Ѡ\M޾YgxRg#A ,qKfGF|I$ ұqRfB}2NʬHvIk^jE6i0H))E|#SF>G= ܊e4%6驊{pJaw#n6tV;(#Z/8JVB.f@1Zv7 X"?l( 35_`j~準eN$yf]4YUȰCOe"B>nN?g7_ȪW Q\ϼ~?AǵTH216GOxlz|usŽ՛Kab^vUkk5y*=?IY)2'"J'h,ŋ}Ew8/de0EE*ْ)b)2xd6I  |d3DAáJcRo7?~G)?z6Bp?S7ubTLYag /mBEB?Ѱ8d濄J sm 9DڴfH+2]/|m4>!z2jXj'_b&Lfin%O3t RB=dɑ-/}㬐͹ϗ'dϠ>k'l<;|RTSݓ>;qg 4,{ qtK;: X^e0Vy|޾`vݙ4^X _N PҝZqïEL yKܨ{5{V@nh_wڭӊ9 ௾˰Tv`_ NJt0S=_5-hҊN$:MSPZv;yPХⰘMF`.4cvl}5;jw'؇Zt>hE4d HKvF9 j~Waoi] bd$'TUFLG1x빠+Sܒbۻ$]e^UhD}ki0ZvO7c|ީ5DA2ՒT"&RS1Ȩ1 >֓ NY* £|L$&J A8> rZ(MIMPU[j'[3AZd;ji U7eV@ZEٜO]~:j:Lk(⦖PJ%S 2%DQ-HJ=}сBfTUK< I}c jԚx靅 džIu=]u`;ja.̰tғWf"ZWzxhb_fb0njwRUVFrz=m Zͩ9|$%cUl$#?쯎q'&q4ܾNcksSɵZ`+@s^ BBuB B#njL %+}q?\lz1c~+f#H)c;8;HˎWዯrxrFIw z2x"pFM.k9e.>"k$BE'3(^&v@YJ*% )(=' ֻhCU~<ɄILԸ~I ڼFA;T:#DHe5 X9Q(EX]m ($]18W/']1E4(|ws{ۢI@L CsXK꧷Z@m XOޒ6#,}W]]LFEe),d5ӭfD"Ph')M]eV- P&[,Q0ӱFxjJ*&U1>ڼݢV>x#s-C. 0)gaaGpL! kVb @U#=q Rdm}'P9HKR d%!dY>)'W'ݪ2o!$QSsUehk%(D)DUlL#-IkZYqv5 n#|R){Z -}AiRBDwnaaѢ1@GFăjal^> $luK`uG^>z*j$Ѵ|$I6Gx)|Y6ZZ|3,5sЂy<»1.Z>'Vy!M]k a#k]rm.zVhՇԖ7n>zʰjtQ5@ny)Cn "$v Sf;A4 #JC]~xӬ6 |oYlxv<]m1?Ӌ-J<{Hkctra@6.. B+lͣf2 8j]qSayGGbP˺'dW.f7iakF~wU?wŢ8/s-zTL?&#::io?U=tӻs'ƧgWcίՇgZZd4R@yۣAiϪ/vyɔ)Iʩ %.hk<暄U9vD lo2,WF%@bP t^@(%bctzLz>H/!'!bF6@c]RRE]RdtJN9m {qVs=dd(@K?bC56BhePVLJb臞f9F7@&C靍m-&Pv**66yW*2&X"b,el1:0+UG؍xr!<}Wo  PK;S=do^bK'.qltWҸ5 ZPl(<+v:mAïd೚ F kqgy±iRI28~mȀ|{_o_NjՋ?O"%H?ks/G!Mva}w[70qM"e#r*!R0ŒVJ ){S y^3_ЏgY~x\5z P{7U|xe]N|.;'('&|bZYE ITtP[9%\R nH삖dua4]wޥ0#j-yC!euJ $ /h4_' ;fiI0}?-ᬋgTvhuo7]D`[I:`! &F0q jS "޿2uKt@YU`LtaІ7f'PFJI2Iv SW@b ii=&A~GWrg'Ry}^: ?[c5$lJ1S$@ͨE0 #= e[!So"%G- 5?MA-[ٽ7HGh{V"[uJPskA;"> kNh-[X1l۠,$  \V;8 88R2"n&?]hhʅ&fܡq\Q]z:8άy=ouxoCXļV5 =8h80 jX1`_,@ccT+&?&'1j]7C"9nlZ+Oqi@H4"Np~UUZqi^f% hB!&eObr),̫)|EA u( l Jx *Zg4E-X {m2CJ"Yx&2F_&?nmw.y \.m?}LF;edגi8Y˗WȺ sE Ksrw^g'[qʐ+٘r=;qV|#i?ƁQ}a58.d?'0ܴe7Q!OݲGΟ: $cӑf*HwQiDPuv%y1nx$}Gsœv9:ĤEcsbKtȳag,s1zq׹.Fהu噗J[:i#9rZ&OQ^:p\wxi/Q4?ͱW2+MF$ʩ߽j{n !lo㕵ȟFSz#_gdLM~.Θ`DxS̴ *YXb.n XC LLphM#WK)u8OHt8ݴR/eFݪcu˳ւHK_gClV*pw~5T- xY+I^$q)z:|UTh T_"]] 5"4V=) >x \eu{3G3EHړ".u)ۦD2%ЇL_撱q8!!  XԶjZ|0Өc/d0$$X}6PLTBðQUs[iM$D fy:tBuJRkܠL_mbFHSEzCYXS ޭ1Tz{2V#Dt=bJ60βXn\͛2Ȕi-͚v;/?&y5ߟuA3]Ǜ h2tBnѻbPtRĻ1V }0Wӻua!smoS[lwDW/#,DAL J39ƈkm#j<8r&pu Ѕ*_<$k5uX,t8Y: *g:Qo{fr@ ~Ҵ =9/";<9%(LY.6 WBc苿\|Z֠ .fk":>7mDrqz^2_7k4#S^F͇KK tH89'Xͅwf@^a&;8 ѐpyFjeaEIyR*ñ!<:^ rN  ))slR:rYU ;.c`!P*fi=p ֙r.p;\K9Sc 2V` e,98%p9[4b2;5# hſobf:ZnjϬP5'*1EXL[~~t;{Oniliq?F /.?CL\wwJL[| _}?5 ̋ ~W0zx X![4%5Tc**!W_}M_p)'Hp"t(DksS8KwK Y81cy xշd2lam ncv[wg1ɰGłIꨓ es0;ŭJ񊃏ʰ3-q< įx(8rV3H}`q|5;^?Z}7|[/x*ڃ55 ?akS-p@$[$J<)L+ ,~$}V0 h-ClɷAZPDG cDQs; 2 b8 q(r 3R32$ý3ԁ`9JLN0S)+s<[ ,TG )y1z ?Mg<:cOn@~wJt_CB,RIeI;$wHHm G :d 9kCBJ8nޔ ˤ$&Ck g%B\H$)SNbz<ղzɜO2:[}¼#_o 0 l+XxQȳND)Lgϳf>fXCk%z!7هXrGf<Eh 0x$ycb>dىtr sΏ-hFxs< _5ռD2ІgrǙe %ȼ aϧټx4suդB ƳWX:(>u!IteWTDx #K$̗zZwU;7A2K( 9p܂#<% l"ĕ@s'w1cPe5GH8zy3IoiyD>8yPk]@HJk_B4qhHpXkC,~.a3y.3VHsv>b*vc޺ֹ%o^gIUp0Y/}\z?^%3k-ۘ$6&1$f~X}1(8`ȫ`4`V?BR7[ |氰aSK\|ND%b(=j{DtݤbUHkIH_M)1on =wlxu cBvޛ!չÎ~H /&T5)b-{P_2ʵS;ө3ij9mf9Y4(hM޵u$BeR}8vdf4j++RAԡ.<IυMّ#եIʑA;.16J)̜qY6>]~{S)F*JrdR̺* (A%q5K5ykM?[w˛I_f-z9HTj=SJMKv9W u%[% 9I] BIN".W SHܤk5U>Uwm"EV:"#e`2%˴DI ZNȷn3Jf/U J`9e_Cb>C,Y2 o¡ J1"pLAv滴U"Ƹhɍ$1Cgh'`]6"k_Kn@su9Fps;̗aпjD+H9b01B3`=A6z#N(ͪ7bC%y_R9kT6 HL>dU;N gjUD Т9: Aj͝]ۑj{u:b5 VN2vbږ+Qj[\\Pg>VxW *UF톃Sjjl vShH |s03JEQ)LZĜƸLk{kK6-GP8ͮULd̺я?΃ƸGGJ~tQձ7BAx*8c4ywZl/53`)U}mBdcֈpdow}FqV,@_H)ۧ=\A1WN[[QFnEQ?m`D#>'g[OmZ̝o5Blnsr !{W&[-uP[b$2ҁq!6w-ӹue e^iZad?C hԲ^zϺtSm;RxbݠGUgU*9=ـn TvT@8" ;CE"]{k(?ߊ( ):c̠^hb7s:gCɯn>> heyչ^SuQQuuQQ Y$3-CGU60s8Z ٓ$f'" k>~V-]'z;qL$ķ]]"gOxyreEJI9J %ܝ`NI>Xyȧ톒O5޾յ7TLG^U~8[1;Ǎ .DS7\9r!|k7mW;`C^Gk:Zͷ]$;J棼CDZ6Ron16wǦeqcܢ7oj80n=Zctn-mu(.c-A(p 0#GM&ݨOOяTy@#ȲxϨQɭB8({ YV90FJdb ǰ);7$"ڪK4 8VrvDG5 Bh$wJy<2]نHV $Mb2}\;h݄TŔլJ_k5[)܎Nkq؊|m^XĩkugV-gD4#W]Ϟ1eyƓhE2{j+~:cik%JMd Zt\a1 Ruzk+ѼlnK2KGPQXh?sVH1c@˜dfťȣFkBcjcuG~c正^kh9R JH *:h)H\J<3 _BO(BNrle/oA\oo3< ?igݓx,!??<}? _/Oo{ٜb&ȣ}BrL\zAz?ݛIyb3 h ^|zrϜK<I2\fy9PXKo&ƪlxD֌) }wZ n;(f!V1A,ݩڑꮹSB[18tlAk_߼/޳U}19E^{EHnAQL -> 3h6d9ZjFUO=]v MF//9{AoT_,m ګ Uj+:]/fiwǏi6q凛'v][bWNK'4LFdn䯲CҜw] Y0ݳ/&=5xēƽyb1&o~krA( &DUo/nϟ'a}0B2B18g=~vsB//Պۆ>mvm{w_C kɷ+۲=P6/ΪXL+#* )! s<1b6R2Z 1>T-rk+#U?dSȯҾ=3qDϭը1ʭ5eVt< XNuk_QF@B(F[FRz=n>wa{iRzrĖ!V=7,J78g N~r~ǓXcDO9Ja=wtQƖxO'O*G^AOso:͚*Em}a$=QxKmw*oنTp70[@1i;7 聸q:)@23ᥭ5V1@@9Ɩy1%slfL Ƙ;ٗۯ"l㘙$FO'U=H8r=۸P" VՊD1nE/ MX*i{\grYn_ѐzLpc])8 FiM&`VT[A4X~R¼>C K ]]$f-:#wDAu][Iӯ7QrBwaYe})K smb^ng.q .x q\J5iWpD_]F 9EIK8!~H Rl9l"f0;C`*HGIl7۳t=euX@]9ڧp,;ĝAFX#\JP˱Uw'jģ)Ul`guV\)BHWk=nC]K+!@A XXq0ʞ6̱ԭ^S6/Τ3(Ձ'" ECL jk"mEYF1Y1ij~ )RůFRw{HS$iDLXb$Rfl]63N9)]mo=) K$>UM݅f:}3{?}]hhFbsg2%Q_<""?vC#ݻf/nczGðFl[c.jjvƱֹ)j]S{R?dAL^#- P j3]HY=NR)nXuTYoki$11LefLܖ7)X)zC~YGhd#q9֓KF\2p8IvߠGXE5M$vݺ8ʀv}?Z[ۉӫ0v-÷|-8RhJo{sݝ_]ǩ>/q7?z̹[rSJSk@!*b䜌*H擲ubV۳B)1,h)2"`'.ږ[b7K2_bQ[rSٻ(Ŗ#{SVױؐÑNpk+$rK);^{!%V^c`CPLN;7``j{4$@Z>iW2e\ܨ! FPh`KS/F%0^Jo e,L[JXnoϓj|V|}ӣl~8W'K",z#W.t*ʅdo4ooM p*sz|,{埠"cuHfX]O!Qj8=V;5HxbbW8p=ً'Aر`n8 Z56) #:ntf0@)1jlO\:E$ 9 z${@<ɦ<[8֋D%6%rFK%*Zb +*V=L { 8)jVڂo{ɼIbnR2 D=;Io)!UVbFRႅ*Pw4hCa_)}?(tj P*)0Gvb*[dGHSH"I9U^J)LK-3 }a Ivt9( S2i$Fr)NQ焨sVOU`JQk &KC0ST2+T@,Di >!,G%xv:+3aӈIh/ge2)\ɘaoyhDtJ|ܾkDt$pHUwLgBVNثSNY>=;>{q9ܶ͝gP;4f˜uκXUu1G1b{퍣8,|uՈBQ e[uPZ~ң;or6\zQX?#EAL#E% Ga# ҫS Ys]bl5~R/hi Lj,9h|`!Ng$SWH$O=O:qPp: &=Cń>E1 hu1cqǮYQG5qBm\h>)Z8˱gAlzEG/yz,:YdO{r#}oVCx9 ˟И%陏Ozr3(͔w+wfg'‚kιΖ0  2s̮Sbȩ|,6ġyMK[gO != G ?_*!;0L;I2AA` S JEAS=YK?XW84ÃL˵uMd:ˏsSOqK>->e_M4w!*1v^ 2C%->)DJ~ QT9o wG"3҆)1m4sgF#1Le  P6Dtia*lJώ}CC+BT`ZfӊVL)!J Kx#涔 =MIGN'BՈ/,̺7h_SHZ'6H*c?m4$(Q+hø b":AAAFH[R)mꒋ f'9)@x'@c~L$AQb Ę#i|5䃚'p+ž`4ihJNiT}9+ #4Xʗ;l(M[#+ҿHM)ji{6 ^y( C~ΖMeH{F]2!uOoUv/RQS&Mx_Wx.Ozn\Я:ZkkXi˻ӊsC\xuuH9=\KO-JAUE-~H3$R#wn1hT bD'][FH"a fh騍w׃zbLzxl(%nM銼 DT"ӉN%˹2(rOkPJ>E;JB|+ 1}~CX]ozA=Ώ6;a$/٥.|h(5Js*/yo{$CL3BSŇW/l`ً8xIl-hL Pj\LS6 (/)?`|3JA_2C M_PKb ld#xrWAC<1SYv0._=~wT#haX(l2g嶍%/Q\4:ЍŔraBUVV%B6F`)T,Xy_%Q-Q*^1Kj|OǪe_hz}NW`K-x9ƽmd8~R &"7ң]֟qX A/y.,7aKY^?V&R<MEPDRgW]5E1~9>6 ]0ɼ kGB9fwnjI C/T b=~K/u` ^XT^͛٧~_`-ןۚM;|9 u9a:&awﯦ[`zR+ ,/Lea Ggg~2qtJ53$ )g6.\ ]Rw,ѭW EXX&),O]]aGX FVx 8^!)Rrp&W1kE "t8de1~tw^onUO츛g]J7WA/ޘ_'$xi*OSzE\)Ľ*@2J^T:0W m MÎ$<%O,VBPum[cXVX21_JO( $ͳTX*]j)=d=KX!xrٙU S3Cixؒ٥LNd !XlQi?;N5 I-U]rE MlQw4qS9 DNƋ͓1e35<Ec+)*!0=i zi6ҧwAP?Q"}O?7r1x|c̐323~QD{mZs8Y5 vfQ*Ps9մ#Yp ZY5(81_Y.j }`lm?f_fa)&5XurT)j8C(M9Zή=aʆ=<*k;NQ焨sRȫ(9Fr^93CZ{OdZ IM'.erL/i2";gnURu.tRs%, iVmƳh"]Yo#ɑ+^A)@? {`iyhFDgzwGT$b.e{R}q{_ A\йp^QZ>wм!d9msWK/Z˫61t~%cj6"}WS:T]A"}٥NzuAruǀWҒ0U :.QܝT띌+j YU=x{mUfv}{CgU̖aՁW{{vq_}V =* HRR)%TR3&Z{`ob3S ' -էHpuR'4缡+TDxi);;% aPzb4/̩3~.e?(xSd2T̥]@CP&ܐ*cd;!CJSR,-y\nt~I6kGjХqغ55$*j)¢#[)1[&)A& 36&Rd4~<*0$Q}ĂSۊ^An)ioikduqi)3e 0G#Bj3w 7AℓYx%`"XKSD* 7QbR2.T $n)S8 _N`JZo9ZxI8@߇Q>߇2ʭT|5WԌYzYtte!-H6Pj`oٍeg벳Y+;ԝmC{.vȽ5?=8bFeL@b`.D\>f$S@6LX@#4$cdP2Dm@ϓ=8NDQXE(udI(9E3r ܧhPR)hRi,q7Rpn[ã" '%Rk6<"Fӥ7Wg4}*\Wu^ட\'ްS8w_g=D/\ XmU#%7oNB?O?=ZC|_!Mmӧgo WF)%dNfvqT(C9dd}J{D̮א U }dxb}s#@kd?x^# 1qCkLK]GE b,__zԒi W3vkܮ}HsJzC:*!Sxy3Zm mcI«*9dB_<0ۛ:Y1/&`d>8Č$6r(OFjZs 4: p%+7KV|6UɊ2 (WHSnG'AN[P g.ϑ8ny4O5@3U18z:p1'GShßgALN ҂BCY^7jpjV!ˌ Zt5]tWITzw, ++S ƴ2SAI҃e'd uX 6UF*K$pB2bMY@?7IBj;bW'-' 'm>*6zmWU:{ly+Sb3_ p0 oz!;tYM[0*Mۜs#,'R1;ަC`hڶs00@f!o1W[ p9)ۦUR}m{G޿+"TΥ3R%! <ً;F^PkwQI2^QU!R*bpcY$@W.PWT-O|KIKcᬈZʀөZ눰5&+s V[1v鐵ñVDbl=8K*Z6X!KTA (~&*볈 TQoVrUQFC+Ht]GoYm`ቄ`'{ѵsch`֨@x$-څ8ӬcG=I~T.mDGkynuc"IdJQNzmLѯ\fo+E갯|JO?.}t{}9vI擳?uUݧ//N"aa/] q] 3' &Cp*")m麻Ӈwe%_]m8+FYw";.&dvܠAQTډ㗼a[e2)Y8[O"kQ͟ݚ܍W2[nB嚲DLh ,$FrX! viġ2)uaE| ؎6Ҥ)5#HG[ŞV.^@WÑ?(*%ZΉW;NC4tqa*%\U3U"5U\VG+I OyE"\8vQHB8bX$Ve#&<@0 UpYԌ ԌD >UEp)FUdt5~;^,eP Bj9Ln&Wm$26@[٣ȷc㠉j=hC=[2$EG"GI #pU9R̓:/2,Gi]*B4uiqg53_ AXoVE#y6>e|;"(> Kc9z8xI7*;&SĥBJGm 'a+@삳V*|3s*K?| hPp:c̖Tf;>s;>s]wR$ǜ4L"K3MF1E( Q Rpn) y NjOݢ8{v`ЋgPΧVtv?vњ m4[9J_FnhcLR[ڙs-:]@+eܩ.1Ջu^Sû%FgyT[ܡK,1g)45b;wqRԳD:9)>9y[1ms.QmA ̰C yî;2t^L >(* ˔vDL4b}J? k?p:i{4%W@sTA eW t3/?S,,UW9$ș8Prq" yS#6Rcmk6m~z:uHsԐb:˳s/#5\fnMOÐoUhFa0”*IZxk4MWm Kl+ȴgc-Z@CM%B&x<3L8Q9mrr!wVO^}gOinl5K,ͰvUvю.^`忍= O)߯ r--;nw‘M`zm?}g7:-8ae>s|hx}.;8vP SeOwz1ӘVݥ3xOR7mm6(mnKU0&JA" 'Y <՞Puaq!aZSl``^e_Y]V)51.y/t RT_ K UqhU/m!:=sLQ[{=6q_z&Ps{=DGuzd]XBd1gF}S,ݤ[AxIB)V1)&c[GL-^*"ϓ$UP.(MmBk_[715+tkDI{V%75eT}:5,Bjp$NfuW\[,Ÿ|^fY==(˜Q"4|[$q*P?D~zL%/}{?JLpkZclLbq\odd!&\F&02_Fp[F $G !1Ǽȸ FFF F1Ku/U;-dQ!ZEHtѧ6mgb52tJ zsmIб" Tts ZikD!Iל[N"ᵬ"5"vlɥHl%` ~`fP빒E)B%@<ͱ 1.Y~6llv假8r"9m^ *K5;,0Ib-Auk춖&ϰ>F?% (dfZ<7L7_[m=K͛Q9˖Cl<*f6[k9,ܕ.fDq8'߯-F(\9fՃ ǔۙf.bOֶԓ@Їr4̦W^ t:ziA&9IA* ӫ×RG2#BsMF#O Yt7cW/qzq3]4#:fҷLR ^ԉz?;FL|÷Ѡ}1ǫw~ ʔﳏ.,;.vtՂ}GKPDG$R*u 3"|3p-XYId~5ܪ,/KvxI~0+.CZۼ7dnkxw\fρ msҝV0-4g9H{#EM)$Z26ںG2w%@ڹR_z:IthlL]ux%U/lJ`»9  F!HZbpQ+{?ebz8REk^r':d -$םʒk!}БeǺTҲcMu m%(;֥җkB)=Y78{DW[Ą!ɫ{tlIykIH!9 4J@BZikr.,-. Dp!ܔU$(,ԉ͚=DeŸ6nU|t ӄK0fJn . g%XČ[?q]ZTHy5iHsbVȪ0,5r eV1B<'D 7R ΍ ey$\蚮 Òןzzwey0_ɣM7[t":GD%t9tZCE[8S2%7dT=t 6J c9A)P5;gjΏqo?NJvq~QD/>*@v/g&|$9 `Wǎxu0ZfS^j }5.{Q ,< PpOJ)"N!6tXɊsl0b(K4Hhb9UHu*9׶@[p#LNOLڧݣCĔ?|s5֋@F1RS"2!T w3) ə.[GF@iKp+&\<`$vЕF0 2S[Jpb&jyQH LJMNl,GUat:ٳM_Lv~B(-wsR C;wN=m+`XRW)6tg,'.Py'Ja@C$qHs@GpBR_7H:2G[/$&4iE'WV7oqRr}`j4!wHTR2vCJP7! %7 #+Km2i\FJen*̸-F|H[X*);TUA٪uƺC%NˀA й`'UƇ;+ $A=QP D +h.tI=B\DyPj7,Uq/݊1=tG\̷ufr[v3?4dzW^ĩn>ݻ7GVf78ntsDI"D 5 -v4H');N"RLtB4{TODi꓌IFdJ\֐{L` P;ݮ:d¥?'2#\PRɰ;${2.@>&Ǫi"C@s82PL{>3fl Ebϔ[QqbKw[w!G}(1lQhrK݃"Ta~`+~V5`7 my0`$ @t`0{gA1wsH+X |ٷMF}y꣉W!5@ҮbGHZMl5QHU dƠ8BK)5+,k2K yd23U/ hb ;0SHƤ걟ƤV/Υ荹?DV`ڻ;F︄KNf wq^ 7/ -hqɭl?)SJRq޵>m$/{]WR۸dKTyغ(E˥Lى, 4~ӏn _! .&- CbBn jP>ٯ<@QwmEhd](Um!nc`:?sԍy¨(%y}%1D]F!{;]ڀPtxv\P-rrdzep.ꨔ߈ րDtu߉H9 J | 11ʎzOٓ5jq SQ5,+Ԍ対7 @ghEO=gf GRٓ%⧿voS"uo!sgEEܻg˹N7SfHB>UO^^Y(8ߌeT u_f3G/в?oDto4Eیʏm3.H^Ϟճ3yLpqupyEyլ'j{4W+ȫaS3^/nbtdV-~y:y^=Yu182\dHD' '4Q2zA ]lZUm&&S)T[{}>T#V<(ku1FkzԨGW%&#z_/k݂HsD{Γ K8sNz>e4iuTfpY{͘5 V -Y4N}4e44]XT*Z|@E,>%p3SbhB43Ʒ cUMˮ"S?6QZ崬!k34ô`y_AQpbAEA]hE,9qECϼ!-ac JuZB 4B)F#Qu(}0>XiBZ+~*`9^r HА{EihA -*>JK@D%P.(`%/ W.#^?xq<^L.J̾qOQPH$Bt80s螢TIqB*P ?bڇ3N6`L _8%iFAPfbtҶ1CUւ6 y"g!_rJM$--B0Rrm? M3F6Z#I8\ SLsX0mTh@% 2% NkPMZ7 cʞ=4B`4_$}@iWp/#:A$KI +FlXyj@8 Kօ>f"I[a%K=D.Lp`*9e ;%#~Scn6qJPCL7xhm'J/(uX/!M[1ci0yp^y_7N8=uʛJb XKia002TG*]KꜼl?,$&4I9(]v0\䀈%Nfl6%Ow{1~MkP/|3F@,CGmg{g8K m%Y9xoiD+N|! #5b6`Ңm`HQpPZ+ 9%}0S>al}6H+]Z(R2=%RPDo O2^ S%7Ӛ`w38R;@5fiInOpJ ,ncpM%ȓX*-h"o4R uf 9&i%a}T9%@bR&[Б"hDTO68/d)J=mP9a&U 5f60&J =9-T(j"n)ݜNm7A)XTQ&8кDwˠ.R~-MJdK|-b@Z"3SAYb6+-bFHl&fh|b\52$Pt:7% PH^3)f WstN'l;&>ɭ{ЌLTCrs,&9gJM3s{fߜLxWcNyzipG-Mg1XSMVOr=Pv2W>(k`WϥkQdCb~䱵:PsUX9t$!_;Tp|2,iFmvSءvk\j$+V2E.0EE |3AV*gL Gp I{[}r{7;ezEA(~OKCLx&&<ρz}ph@U8J+ן>׏%e l22>>n+8cmdf[~Ue6Q|~G ԃsG> mN/>OdǗd Tely8l܊77IuU1EjJǎ!OLD]҄Z]IuL7!=ge օx #{9޻$ Fe,f봹Jr:N65JypN*eF`ҴF/K#ǵXi^eWu=Bȏ[wp+m3\.h?>(>58vv6gzdi`3C-zz wz0g_-Țѷy|U;2jGb^/j%鼽Wt}NZfw&_\(s{sPeqD&aw>eJ`딝pZ&ea2شft}{wvq(wh/g'oˋofmxx٢Mjs9&0/ |$Թ@  QZEc{qTf&O ju Uhj!%КJIt[o$5E` ԡ県LBj QDQ2n 3QhPܘ V6=ʮnmAI3>VG;*WC(PpA~j!Jأq 6JS@ƐY.:q۠"mR3*S^TYV6&G᪺iݎ^4JqUtQgbȨ/P6P&_}㩆-emԈdDŽ>Qd8\T"ݐXXjjw%jHl۽aIgqy90 g gJnT4jzPMmmV .|p'pHfwUB_Pί9Uթא +uWk-iW2J|镝6c p #q{َni%eTۥS/BPLFD J7y{F`_N^tZa"!]qBi#DbUظЈL*Dm$1$D2$q21m*ڰq޶2t]c(P*^(.4B*A*DWy1҆UIʑ4Lɶ6G]IΑ<-XّȞȶY=fLJTg$ST@`H-taH`2ԏ7@LX.B <цJQ[9LjǒSI/G/[_N)l'X{攀 5%{ MK\5 VRkw}?M i6q˹97Ws.\m֬ca(3"J,5g.^U#EUv]/~ `Oͫ3N +JI*?#%2//'`EZxUf, wYXsʆu&fX}6Ҿ_}Lڮa(40"mڟ{މ`C(}d[i{vyt(i vu+cR N`beљ0D Ϸ ) T~Eѫ)8L_Wl`;cNsZQ BIZsf]!i@Ne-W3ldK3D5Wnype]˯ dQ3Kl S* utKjQRd^t[{B-z ?#/kړލQA߸Oo Fes @Ռ _[,Jsh5,x3G'Gu sry^K_#>͚5GŗwX$ja4ػ#C[hQ k-B;VXJJ,TB2 kKOGQQ vuQ.{u9Qc Xւ~̅ġY01s*"NBȒ0.$o oꉅ[2F}U*! cÑ3鋱2֍K8 (JKZjAZscH!0$%|2BM!cʐF)8ה X2 ==1i6< L̂JZJRޗbP5OqL} n/ VkBLD6*Hk=v(xc٬lG" d+J @ٲ=Vpul.v8A ʔ@JEawR"IRPR[hXxwtz ZwSq;MogƱw0)KDi\2҄$$2HL2lmԹk$`ZJ:̙;ˌVwƙuzO',wrx"[Mk'gCNSe2JF_gQ[O5Sp?@NNQJ Qu_WO 4pOz Gj0fZK'z-Q4jt\IޭRn=ƥĺѤdmC!۩֚BX7fj>k {1_.'A҇uwÜj~ oOX6B-f*M(fr slrkQM"/)D8J6lFìd@VE^R'ˊ.-Bo".Qr(RL<Ţv2=Ҡ漷LPJ/-b,z~`- &j;6Ho9jS Ee&6#H&Rh%b5)XiAazMԷI@>׋I9)A&+iyA\[טb'*nXJJޢƋ* 8Z:FBN0Ea᪑␊Ƒm/DA<3x*CRT^jʙt.XQ5z20KsE__eXfsmC(~'gmƜC=&*7혳NNPSd9 =n[?U %@u#![ؘz)"Xd#K-AP\feŨ )/ ,Bx({C¥"^Lg \ )etԽ EBg ESZjA-)ˍfSeK_ܗcj5N oLa؀a  vi) n:ԒR[ %l8JBV#$4< iB 1k <1٤4:dbZ+RK`+ƴt̬c8И*.qWKf UIdY=k8$E>7L/Q꘨1+Q3 `B* 5! "u V(zT Uz"spz*a(m4Vei)hHye`FV}8k00HKm^L%iRϒֲr'g_s3gXޒn2Δ9t urǨ#R 4լ7閽T4%KugT1{V\JD%%RRCNBSP{'Ob]b ' ~1˻ Q9NB /!R- ec[T2YjJ4}XhK5hRpȣPBp+ XTsE]a'tly-^_;˟^ֺZSOwӃ @'Փ*^%A]um77!U>hfS-Λyڇ'}wR!)*-%esM)lV`gXk0QSû/Vu 2'ȫ_ft130}XCRh{*#L9|c{$K)ӋueAWggq@WsFj )E"mf⨐RYM0Ž[휢PS sט=%ЦOK`UJ|YRқPft=9Gn2,>}1ˉ=[ }Vz'9Y|Z6/V-/>cOyþœœSD8ڣL; 7L V99QL%RJr18QV72 r;@*y~2jUe󓆥j̖⽧A*oMڦp)eJJ K%yQ0cTxDev[ ˡ-傴Gg&`)NYqe F$5hBs d@JPQ`cQCdFK D8J  !J8 R"BLGɥ5eS%FLWZW"%%(,0,TXk9q|$sq{[(3-se?`lm(:WOm@K>%wϫp Zo7w>_5/`hW7 cDƱ7(ɨ.z/Υ/F~z57X ]Xoo X u2,O?'w%֔COاn V`r.pjEZLC\f5;^J?ɖDnlAIp@Ni:f.8dp;% K]HK+y %%p3y hi^:.Aƀ `Dh{Abu %Ըk8|_kZlLMD$ X+רGiuen31#3yq.ܸܻ&sw77h'<Ոc􂥟|vf'ӳgx?l:=uHO_H1R8Q%K_E ˙~޳rKӔh{?\Q;D90L[7hT$d6p6 t\.BJn<3 3I%Z̍=9lҶ0#lEZr,b%b%Ok154S$δӫuSoFq@i7`";aŢHJzCa0X>i2V!8&6=śٛ]o̩p㋊T7 D;NG?Mow3;P9~Tlj\“)qwMцb~/gz']L}zﺻ^S0J:Lij쏺SБa>0bo"NB:clY F¬wRxOyO,JY2AZ}[n7㓹[ Y@Ir*@# #/]@^qemP% V Rp2vqlaboG$uK"Gj>QqRW4# @d)<-Aŭ)%e2CV8PJVʐJj~i:9na_gɋ͘Sp$?]{C"mkw.5' d_]uͺmwр\IFo/Mб+R.4Iq2ag=;Oo\/8î%_1j W3+T?c&yU|GO&~&⥾'..y@?+&Xt6TD.ߏQUU* ʹgJ1y' -iWY:{UuŠĺ0:N ݛu/VӺ!_fT3u˾߲ b^r>; 58"r+RMի. r)\3f a}n?O%Ś'w" /2@V&!Bvpog7$5v[0f<wȿ }Yd~P2ɿ %$[͈I{F̅%'0'if,%prD %P{ј}o :qo[*oqxRKoنųZ.-唷 elT ,AwO03tP̨Ҳj@-v/]%dŏs1oO^n5mNap8)R0xQB E%5.0B $΃77\\@M8:'Ym-to<6#_b_0Z6AZF`Ub`ŇņopEhAucsu_D4%Bu9^D38.(#@ODm7ͭin'ldJ$;Xmu?yN&]TSS&@47!vN}0FLa^jRxAKḒ gqfJ K% |$E#:\ e(1HJجifsbq0DiE iqkH,'AY+u +i?U'kBf)]LZAҭ\g(ԑkڴ>kcֵ?Wn>޶zD_Jy5s] *L 5 P'%[^IKݩN$nl4Vk1X^L)=xYJKW5ALrPënVgkT,cj.U(NjIO2 bH- YXExBXQ$V &sn*d t kQ53,1Kb#L?!ņ:q x"pe;=m 2Dj90YfM-g;~TqM( \M>=#n 3%)Vx$ICՇR/9) EpN1R\Uz[1x6*JDd)^bv(Q \zN>hw$k5qh3.8.=ZrJƻ g Yr*aU8g)wケ覙A8q{˧Ǐeqoe}AsLMl׆ I+%.~Mҧ3f2+2O'S~d0(V>ϕ_ƚ%w~r?`!<$=iv@nUm,e8j8rkS>3|rKtk .l+6AlYuo :AcAIpyGE0>mEXHv9جLZ֢ ٢SюWӤO$5c_ֳ7>G51K{1uf16td5Ɵ{Y֚1}" 䐲Ͳ35$\;g!ᾎ5D;qN|q9< Knt?} kfhӷE ~4R:l XN݀1,*XŰ s߯z7\wa<0 &01 }/D,T7U@>gP'~^I #ᩲ䜺s} U^p<& b & b M ?En~U8UjBů^pWyIXRи9Ӽt_.w9tكhH1js1jsiU%AjIpYZ;xxZbٖ|@ Lb_FjK$ƮDs-9bƜA"d !fe5Tj%ghI/ũ!1L!ʹqb 0RneA/[ӦZ9\d%"y|z'~C-.KKr_Z+5eqQ| 9@3>Lk8" 77@o8O+$x}!6{(%ŊGoUyML XPqK{yQOvty7s۩pO۹N;yPȦ' -G=I:ǔ{GVd˝JU=kO4^ZUٳyr Dv,;գ^(^[ॺUGmbvBswukal2/ K\woF[_T~Nh6og_2sku3}~ݲ~=f *͞bn Fk&}/;btǕ3{.|aI˖Di`ko!mnE=S!p䬅oI+&'z]a4DD}8V})˖%$XlsSv%dcfL_?P*qs !8=7~y=GBA$t w<\E j0aK+|/ 0dȱ+%:&7w׏A1"[V_BlUj؝:҃SN㠌%KRIx&f[m$P:JO@Ym&8,p/%rWDp\b"HUl&¬J8{(`) ݗ_ a s~xˎT|4P/rUw*LHQ9Lk*%3D3-6a'lhmjYecDV6ؗ@F1xKIn|k a.>|V"blԶD#b0 lWcSRhS0)C c\Ŧj}<UrQkO#:Q B=&EmL<=ĩgWY18m(d+acxyg-Nt39Enq2;?qb"!KbN$׼su!JI:GZpH+-VD.9jȇB]٦sԀVꝣU19ڂ7fu6 Զޟ=m}9݆>%S;:\(hr>eMeR7\ݧ))Ƹug4qB3#xv֝$`䕰8 H6ǔsYRlԎFz˥2$QvJ+Fbja-^*a) !tLxy"t,nA̖el X X*΅(L'<?Pc8%A[X*Yẍ́\[r`f w-עjQF!fK z\bn4$eeJ*0%żֶRV܉(o>xh8sk3$0) W)AT//@@RY6Iecz{ )KmۀfTiK-/J7RJx*""GEWM+~ܷo&c5?LO'S~Q@>>c-`]B?p3NpY5[gSJSl@-^qIJ-t4HQc1, ] !rv5&jذgj!6 J@5[Om T3>$ q0c y/gv P(qROfHٻ6ndȸUl+Nv?T3֋M "S\rّ84~h4}0xVf 9rR.j_tuqEDX9,eQ#P)=#oHl̰09RETXmnS]w{'HT; *Ǝd5UvǶ ND;7&R0hnqJXJ(XG v}c[jNKg-EM=tg])tFׂݹrŀsAᠩ1'7z:~ZpR@}Sǰt tM'XG#:$7gO&ڷWoثgzةWwEK-фr⮡̂GT!d_c+U9K@h}y$7# k,,K'EP`B F{j hb`y`𣻾07@5 L8rŰ"e,.f4-nZ[Q=(A(Ve<9|>ɟBc39>՜0̵*?tt_j~uizE *c h5;f`^ *wqWɆo<~drz;XHm_8ƛĜ ޵C45Rwv='1&/cLuҷYҊYڐQϋH0O|Zi6W "3fvI~f&s "{[cKaxTʙ M`{2g|a mيW!)T2[2d&P*ɾu5u$ (t.LfMvEd췛dIX>W7ȟ }'@SwɯwwӉoc|~r?Ap^ھ R9ڼ2o_GˠNqXq+_~!V.ɚ{ש׆]r9GQ P~s>~uuYZ 'G}{s$r!wIO*|% d Ɣp]oXE*4qDqZ~jTR!cM *0&J@?7ȥ<c,/~ɧB&tF3=N+ЭM O  =7mFź$iֱ`yάWB;-)Ry H?|>Rgycַ<^iԁFd1ޏ}l r {ӥM8S(Y=u+*2 ǍJ󒍴Ri,MJ 4"@$@ S DfnR#/">zeL}]nmEE/Do}x>^:1𳹽`<.j;m.uW@_1O̿޺W`M:AH2̬7`nh9`݂q3mZtӀ5#!\Dd*bFDOG햊A褎QEO*Fmkj&$䙋hLQMQs[1(5RMu%+_]r+z:J38RE/?,<0h؋kw=Ր\*;zf53"H3]3oOHUP~;r+vf<~o=ܤpP7ٌĻd&&ѹD@JtҵZ-yy3u*Qɨn/؈BwZ a7fL&VLD#KͰ*A4b{vƸ7 &S;B6텮^M20&MBaeQXMfj`Ej*ĞwFI+gۇ ht5種9Ga_5JWRqGCyCp=7ceuX!)YZ[ ],iG%^l9 ѡz?^5͗z~¼E/\xїu,bEORKl\ %Z.]bҊ7lvJ} P.<.<.<.B*TJJX8 7+%b&*J xaqVURJVI<^>~ؔ&6o㉂$ GU_*r*`EѸ9ҍъ)VQ"5ṓJd [\pNFKbQ-UТ Zl h OF3hlk;Y=|rӷ݋Ӹ+mn J[&ՠIf.|v1 wO{Z.af wϟ 0%o<|~<05`?<_7'`:'Ӆ>ٟ;WdXHwnxs{w A=> I,=O懩w/]T(EE"ܷK3 ;arV-ۿ z/ EkAsGCӕ]q< at[3G$D9 ʄrjm3"+3)HRw93~\+x(/<}2ON()JvVc:HrHn7EUf`T/ Zb!kJ)ym*Xשb:^etS.6G2\κ"[DgC j}Ο^t|k:'~8$ RU<{U@0$yi61F'x;=HW] GSEcwdp[;Y{Wg;_aBxTy66hKJAkm])mg.)2E![*1:Fxwݒ'ݚg.e `C?ö7|ޏ woc D8lZ|9$^MWW>W Xb[sx8TVrh[׈5>ZFa9' k~|whJI>KWJ}TI_ӥ k*OM(SQXi^ťm;BxMW*ηw#2X}2=1d>ɻ&ky[D`ɼjNT e&4;կa;hB%?,P6AH1]F9G+,32a+74vώ3)O2@l BClUgHa:]L!/k(ʫ7waV8;,^OMq,eଶӽ3Vڢ8J0RZi5ߏ:=?BR~iw4B^zxC!Ecp=vu:{[u}@*[@Lg8\_Ʉy*c#DwAg}:x*W5T|rj q*Wf%rmc'&~@+5n\EM/BJm!5`)Wp-mh{SطW=Hgbu\GiW˧_6ʻibS9?QRkm {l勥Sg@6#=J$"*"(o=^+W6ÜE,ɸ/b)~nBګF܍>݊tHYFєJ\FklMT$Ҋ%&U+ql ̴A,CЀ,y{N0URgwi&4܃xne X$pac ]JESX[.Ѫ3IsM@m9rKldZs?~+mVL%3w8g o}p~;FNpPy,N ?e܌` Yqn`CW5>rsw/V`GyƊsK/e7Zx;I>_0JfZ66G!F pmox kc~GHVD&HAtJZ/-*(5AE߬Щ]jݲ`- /nY5hiFK#:b̬q$ C=DWb ܗ16B`%YoQ C_£Bl\(ɐ1(ɰ$c '~6/Ecy;BCBa9 }xZ~gy! (jxqAc$Ҍ (#l!b+Q5V~e f&2D]=\uYXv1졌4$P;GEL]"p-@ӽn9mGpTԹnޚ;* H@$uWؘ׿7@b/ -R5eSJ&0E2r)Q\Z2i7D[QY36`1K46c2ۃZ8&̚G}򮛌4 +.*X|̬"o |)2 nI:;&TeNgtgz)8yaސwd~jM SbÇjߧ2У)!#G5rzQأ h#j:P)> Ձ,U?<^1K =ҁ2'LyFW eWYSCE%Ÿ@^PwH}2OO, dm+Put7/ lPi >"ת.G0b1<{gr u\lh<M-^^c l +7#A3x9#%Э"p/n tmH+$jWYlĺqPX&^qǟrVbPH̵frOڛ9$bXiXThK`RI{{k6x*ϣw?\?' G1u.=\؅b&e:ʞe:[,/]'YR*ʕ~w}nd @xiD?|u7;{w˵)EvNI!i1W`8Q%2~%Jtк?~32d` #صc,p?2p.HBlTG %Y*CdmqߏsGĄw\;B5?Vr^ȹAf6 ;M^3O/?N^S vQ=$< mDߌ )-C@׭W%ep,S!x$nSyX:O^bĸl%Y?g;]+ 튇/JT.wc~!Wuqm&/waWy ۾?,9Qlx%>v"Ҕ  Ba_ pM|9ɕz8pҚ 'aQ&+M`bǺpq50nPoOuRo$cz ^m,(p8B|;XCbJwԱ[:CAĞI4SRfyfErbp%a89љN8su`Ɗ #)? .JQ܍/'P-W h-Qcs")X)#Y5J*rW92,9#&Hq*lfA4$\L3 D e&W& 7P$)Ma!LQY&r0YCXnViJ* .2hTa,iP(Ty H:!k$+KJɐ0ls),e 7uKrM@=Tr LxIccuC8]B!ɮ8h:s_f[-DMgWri/Csr- H{CNjźy'33_?-2 neBPsB_!J-'{` >rmEOW1FdG? 4|l.VE+F߹O˿l4yw7yG` mh"t]%؉Z2,0l _!x-j4M Gxi M)[.&iNj"5zOn3<2D!p1%I 3Gggau$3yG 3tC,xC g.;6No??Si4`Pna+¹X]x ZjaYn`4SA{ q)u>xspH^pH!q3A!"ͱH x_& Pۚ)y(Aؖ>deMc܁ڡ$:o>F01_[`4성 zQLۚF큝h,EַbsO.dUI;aLtn;ctfQ<,n'XQ%}Ln'͜:kW ~ 'Pg%M9p0\- ̧ceJ8#e0SHIDj ϹaZ>T Ҍ@~3z??uFM_t3rVC N܅m ,ww{n4J.=T!!U1&,141DXCH(Ͳ$"!<ՙDRcIC0샆eV} {eUSjJ֙$$/1B2)pb1Ms)3\WK\ؔ56ѬbC,e"S!sHL8e\s 2JQMSH{*Rò*kJ0!wυ*haBR eDb"J0W|R٘Sx4(5t…G'0~ҵ8K8K|~Hdmgc3˟jZ<%`bu7͗'ׇoSX+)*9h=O_9A/QR;/fM<^>Q4#aH Z-\WbIq}mt` u-Z3J<1Œ%SR'1cf P"~P*t ?8U A8ۉq,r;dK$ lHNe/)<42Gd 촔j!MrN1'`? BO7UZԕ[{(`'HbD~&PñlP0HUwZbP6  "1|&K*EY_s ޕVުC>2$4~ Y骥STYd1 HR:"*!**5*l|ͰC8'5ѧ ̱kÇ3#cм|X"B_NO_=o\1DPz@Lvmn:J I!}%,")(.aդs$T(D1,p),r[uOC,~U[ p'I;iO0_|(gΗC"+mLGXy,]W#-(+ z_+FhMБ)fB u@XAQez2>vUҕa h%E ުR9bBR:`XKͩГ= ҘaPJkOn׿g6Y( I/ASL#xTѩ~[ T*@E5&D01[KK*yE& cK|2HO%XY)y#YP@m {,71m[ r[HH *uL & j0=D ɑsI <=$va"rYOgu8\~d!'tp zF2KQc(JZ[`: 7& Lw5[t.DIN5a&u`i:5SH‡`ؗ:Z! N3Bu(&Q`aq+}xa8R@wZ5m/ĕdHf']S(A-*>P/D#+WI ^V9L s`riP͵7]C`{QxZB?_j;b'nq#.CCcIAfK/32>\0I9:+T?+}=n{7A+O՟K(d[k%M[%P6lYmSV* ˩mϊ/|IInsj(!hE".!OlZ|Z|[7JM;dklm?}"|]?F'tyWN_ *hF?>=>'lF)CΑȸ<3Y?r!ٝ?E`B;4z Kas>?IEPO|-o~zuJɛbT0=8cMIߏN뇏n iCn7WOځ #g2[!P7DP L1ʲ%הDsΦ|:$eqn:1!pZCx smzogxPm[ rԝ>KwD[hKQzm9Fio!Q.RtJp,<"'0I-ѓjHH_6"Sk[˴m'gh85FJ2k  z)B7XI4ɦ!0PD̨6ζcՀ '^fbg`b<(iօTi_[&B׵ʷ  +շ^rDA#ZWY2O]蟑d|˝N7&w^q|?yHqa\=(N4'V4$m܃2 ~ J>S?ֹ|ˏEql|Bu\sWCrGyuێڢ.PWV5GW/޸zsK +:|]~&CAY`^z[F.A6s[6Âb~R0MqijǔcСߣJJB3oƍHX@ZLaTe_q>+hپEw V{6?[=9o u8}CNpN/o}QK)H$vXDY_" i(X5跷wE]Jۻ Jw ;Я]5xyW?~v 2t#)-Ձx2D4+UeCtRP JKRV!ɤvVD.+.`%EhH !*֚e03T:Y",r#uu3N$RH$$Eշ4#LUyaR(Ͻ5 dazSi-U`)=LJzPFߢjF)HdQQFVϠz(]:,d;jѪRB3Y gRqP\%,U{ɜFjx .^cDf7_c5F:B.DR68tŜr9pX5K!4 Qbk8t3A9f,XIҫK/m{)N1B$Cq!MkDrݛ?S3.l\h~ }{~,;rl泱nx(^>ʳX ,|Oޓx|{uBP󊄇yOO{(:l؇2D~:7?}ۛݓ}}.1?`x #ϫ_brQ}y2 m຅w}Cy^@޼dǨiԲcKj Ýx",u(`\j}?xa4 oQ#P{$o13Gu՜9G,bi?9DH1QDln>CXaz 7n~s~maխ=?[e$$m sk$ZT3۲R nXPN(..u Z)iŰlUaV#ʸ&%wBGA:ܗkjwSx7q@~|v7\hI)ܵL]س:߰4w!rwtZ^b*YC$wu㿯N !AR'jWqhq9L1I;D#Dwu孢X"0򾪔 u덫-}+TV"6tpP\+k uAm(V,>CɍoB~];. 9"Gz|}`/c 0G*0{vI[ãSjEzٞD*V1͠ ]u_`B%ftbgxϙr۞Au:LrZWm/ju&5?>$}6܎flwx8(.sv %8m*RђYGZyh9iY;_hXVp!ddEApphOHߛ;ZW&M\lPheuJ( FΖqmBKb\ aUiU4NP-ՎԢӻ.; -CVK/>`A|]euˌwczU-,Z4~w??)!d=qRH{5[SwDl' 6P1i6wmdǶ?~]VuHcm]oGsuu#Lf=<] -(2$%gI JIcY!)&jCBVΛSJ]PccLS%ީfUtG`k0C)m&PvPA$RMK_6 V})K,`e'vFYXƔ!T㭐-\WNMϟ?9^ 0)`B & ,)Tt|㕃=1A܏ǣ䬍("Z7:?SՀ-13Q}u=s jR(&NW@~G%Sۏk 83$bQ<Ů7dͳ(7Ϧ=AHY'sqෞ]^`>轻oF8ݜ<Pfs3upr^.qG=|!^*QTe.0.ևBIٸ Խ7Gt{(Iǣ0u?"Tv޹2/4TAz]dDqF9V] [d< I$O,B8S}eйѾ)5߷B+â Mqv~;:D=3%ϲ[6BdN= Ȟ=[>֜-(B UA:nO~P(v*Hjsw.DAPUBrrٗ<9WJA%k$]̊ fVp8LasT!ҋ!PibxGc iD)D-][oG+_v3-b@ ;Yqy@bHJOUݔԼW\20bluSNUKap.>M^86` 5'; 4+' `nq~iXXxޔI0E>7_^PDr֌dcc 1k^\^VfnmĻ쇜 H6ӗ8WR 0zָR[kֺċrY]f# YlAS-CC;kxdQvn=66C[k ) -A://23C3ҙ86K??_+Tuh%4$( TaVA+Xb@*cRN"LCL?^P BJoKg쓝,9nV1h KQ0[ې:~TAwԬV;heB!! \( K QR +{յqRzeE=^jIFb*N4+RM]4ZsIJ`zm2 RS a*e nCBQ(?~MuǼ zJj}Qkq0^yUbW}=:Y(E;4pyn_zfp-ѻ&'r~Gl<#Ϲ,g(P9կ 9sS֍ swhݪb:UQƺu NPk֭Z:4UxbPrOE?w? ?m]Bٱ&G+~{ۅxR.Om 878vO+RSl;\}MKiRt%Q%"žVe@TaogRBZwoUc֎fk]FbW t`meU:W@ JebP '2d U ǫ17Z,ɚ%oۄ(F"ÀNDQGr 4t䫶ޓpF"q`wirܓ 6-QyCQl: *@؝xxϖ+j5N qw=O_y  p:M=Cّ@tjnu_Lvn<U'>m@*{&{2yJP+.x&D*qnK]ž?>#O^c%@画ѦgSާJCq 퐃\֞>A8n]o]Rtx_)ɠLX9 Io`{%}(x. 8CYsqCeQJ=E^ =S+H? ݝ62eɞ:uVxy^|W1wR1іRCjU矍jv@S92"Xm<WL(zPKJÐ\H T![L 6ru^A#Aq a  dH J>yT 1ٺNhdbe ȻKg_Q$xƩDD5盞:ń-iͬl}‰SD`lOs.'yZ7;k7?qb(@-O I>12irLN|pF 7^c@\_9 u 4G!F@mr8p'x{Y`q}A\Эh/1g;Ǣ/(Da\+0= c!MpRbl#y w.]Pcq_i92E3gk)T[h 8`/WLiUt,nUŭ FBYJ2PHn!2B@1AV 4V(#@F${$tk (E@Qz("U^=8D+T :(t.F5}ء>̕1< BncU5UU#/שƻKńqbթ%R;L/P)ꈂ\Si@9utt8 bTNY(BKMJ,ER#520TyKo8ahwlѺRI/8n)FD HL֤otyD}=BjCuj~+'ee7C~dFҋߣL! jُ^;뒿&pƅ* Yss("3,n; kniN3`Zj{젟HC0!J8&;YW GǼ>.4#&_ VA`3b3[ qll~ma)js&2H{6ʏ{V_nJQ\;v;-*vD%1g'E fqdJ.' IHڧE' ĄܜI?l49وPc˹gѲg92(ཊ<A1g=\>ϑ@!K9 ]mH+ц N⁋r/[QP {Nr{rLp ::Z ֗ܖs "؉,>gʭŧ}͵fot(rCU`=xvXc!޵Gɡ pW^pc OSSP;kFUGP#ϚżIc12>ɘR&<< 1XQ>|lw\ ;˟qjudmRY"#sU_'u9vq55|f *?ed~QGrJ&M3z[gxzyzy)KV=] kBt%e,꒩,YJ4]Y%! eendeF!v(2TBNE9 N1,b2C$gKPX&z.: eup6ya??OsϪˢq_+cz=nq(5}Pބ;>7)]"ٝ +`d+W\N?~ݗ-sq/%?EʼJMr*ZI8Ĺipa}ݡuAT}F֡3Y/kА3WuJ HQe۟pL: GajC %ሬNJMRB* U)Wt$W,㪫1nWW~ w!.D?E܀T Ȅ&TTN4ui%CP(dVbSfC]]*UW՘!9]n晞`ߎ 8<̦_~Vnzv3~R(^>:L^>x{TrԿx{hϟR%#F[!rXDSb&5 +RHP 3zP١ Q: ˗nfuU^dh,\i=}a%,M#PbJ$ V!e:d X"2,01N43#QPTjaVpMVeRa^pA|oBm@"^+?X-7 ZS#!1F[}@K+*}S9)ZFK~o+n3%+f[Ђ1QrQ8$B]&zIoKRz`(8PhvA YAB3ĵLc-JFc+4tzY7rTRK0&@*^JfRRj N!HW:"U~I pF!$96`L0r]d! eվ ;@4)&(jf.#btm.J5#OOws^MO EP𿅟Bgᓕy#{*'/_+oDMtlK2mo^* °Z2S(1п6~TQ7qqH$u[~V\Ko8K a~!i" e "IHv<9:(ak+M.-@Dh~VCF#:;5}GДxcΏacx4 wMϋ=H礥旹9j=eRTo,LREQȠNꥈX7Ey#*~)%o(<$CKaޕ5$翂ЋTW9wb&OrL)H.a@j􁦄 5*++&(mz-TQ&v&?ƴOts2fЈmnݎ~l4 I%dI(1\I#F!!){aqh|pI3kH0i, tI''^#F(ljzf46m:zAR2 ʦ))?HEFω&S{"RF&ڨA(#!>N5!8bDA5Ps՗N&5E8tи5ru)<ff  1%֙i`k>5Q)a c_:B*m. @Rk! KwZAl$r^̽ `{1)B%i_ay=)ھy T(FĘ 0UVpfd@T80n adZX%fD4 9n՚H'{2>֞A!7 5UtOS%m  J%.n%B347WbE!@']xRHI GrԁHo21Yj!p*`<\69Bl~O4b$Qn`OٌR if!hfJVZU(Qb_eGi8?Q4Uuu;9YMVCGx_{L(QO 6Mg2N2Cd HŷҲFР0PŌ$5K\ B>( C_Wjq7zk;w$^o _wD+Mrrֻ&PtPQ\R?Co3cJ}{yo؏wkΐօyrkη-{3_S3Jk}(nyzp&V8^>5>.pl۽zhC>UJ;GgV G9ws ]d.gҤ]]cFWT zGgVCx!J|Ƹ~5sk OS? Ν7|onW͹&夭gjr ^c8eg'*"_> ~uQHM&61 $|Z*'sZ7d,"6>δ֣ixԐvN5CdsGҫU:krx6:q4ʸi$VN0u 4RMH AXH`Tj}tQpa毩GTO f E/;dLE (HZ\4Y$T󴜹0k8btI.Ư}rw_ʣcKw )QMͨHIǸ 2 ؐ)-) zG J-k8FE`*H<_%Z(V01@W 8"1.P-AvVsm6Fk>{ӆ 0| |GPB?4P%OS}s"҉ '-\&А8hTPN)8ʽ_מCӍकS}T]ɉʁ1o4⠡ w{Z}Epy^2E*kҍd]49}j(U+.Uxv*]qʘ'*>E O&IkSN.hFc)JiDc*'[PS-j}gX^i^5&)#6n"S%*?s 8 jx Yµ7}sc|UeE|<FxgȀ*x{%xZ.!3%5՗]eo_|d:,z ;i 两[4j.>ݵfIsXNkݾ738y*#O:q9T\ho?NS/-W㗻0_H@ %;"FB~5*kђk% d@d]q1[*~pyroZ1;^7 ;lKʫ]t+JAO\F0wf;ss^B0qP fF(+Wh'8?7q=c݈XTT"M$tt*Oh01HKHJcbT2JuwjfMR~=S.P ~#1&Qb5z*Wƒk{iCmPz]Bdʪk"0c8 04 ,D+ ^An4TVD(NNr߈r9ɳ6!/Đ$#UEؖq[PWj:!QIN_Fz)c)Ϙs)FBR9GkG`%m0%)X|"ӆSxJVD 5nh\(rSOq:Sp)e5DBƃ7 axY>qFtΎUec A:X6T_bT%=m/߽ jhDdʷڲS׫O-\R2#hwZ[$;^[a@i:C]r(-Q? C~~ˁygZ߾:ǥH ʹ7W@PԢfB ^/g X?8*../#1ﴥiԛō{ ⽼ˌɓipÎ޲>Z]}KFF?W|x-|._f\+6Mm] р 8s٩BPBVt|_NI sT/(JкR3t`[ɧLںР3#؎zW};bЏ12SSM廎QNQy/?Y%̖|b'q'QAI-*ga3֓|jn~0:'LNz")K 4#Z})I9F؟!'6B%ç}wOl mWᢇu޽3~p3nIj="_ܞ}\Y>Mvo B\z_=stƪwSھcwU<#\-WqH|9m.ݱ{\7 #|:ec~گ{6ҳ +7$V2U90Vo4QXNw4n[wiaκ>Zj:o\D[_qY7 `bP":mĺ3Qiطuk_Thݺo\Dk2UxD6~/vNfq:,=wݧyp$O޼.^w'_xڍ!2vdatPDDwōuRĉA1UD$\N0|ZP-9uK,2xiX7ށ6 "@e(hŢ5"TZ X%\\27? drPxʁ7 +".?W_?r09ZLI3 OW unMD^ܛT<wK.ņƢWLak.U.d>lG\"__s6<]0+qXp<)S^ D0xX]|MdIN/ada8Y~ooM⣔}xV֛|OR>rk۔&,.}# rg%ro |5~B3i%ko-+5bG:ՠHRS(fF1~%'E%YE/kCB9mJH-V89jA8%R!c-HՂ̄fZy!ᢢrPX]< EvWrA#ШUBO}p_Jkv6i0W\?|Yוi X"YˍV,XR=9KN>B0EW LdPUp8cz1]RS:`39(;܉+b5ܽA* B4:iVopJ^l݃H!@QSĜG^9[==<_ʋ΋Xt_ry4㹍[T"f;y.ќ7|za޵ƕ%/E DO(h?(Lqbmfq/=>CEš"}álWN)Mbf}y-e"R֞#s@6m&|y)>Zʊ#乏T#&IMSbv֥T٪T/|x)PX,ayQ FVd 4Z}Yp BI̢oZ (*0LVͺcHa56l VȄIfRfJ&G0 3$f퉲٨YXM:Fmď^,@߀Q\L#͚9H~Yt. ҖDR:B;0"Ax"C>hx+ ȸgLKdjAdiAܜ7k@-A.1 ,ʥNCenmkV, (e#eȪ.":ZނY1_#)me[JO9$edž2?4oā>k.2qJ˹]RVTDx9s#e|䅹3X$;gKumh4#JƙmG"YNxIFCsZ@?Ƶ㼴TM/OZr8/íI3ťxIp^zIc/e߄A~+Yt뜍9jąY8/˭Yp^zχ2Mm= 1JT;%*%| Agf?ML+ N~<:܏VaRE?jr(գs#&T=%'΃2N }*F u+.\Be*Y̺/EknShwtʸ8t:u˃>Q>mF6|*S%ƬMþ=}i O 'O|ovf4|7GԚ>hiRmŋ>k-Ԧx#QEh)Q^Ykirb}nM(ЪI&;v[)! +X"/,dw0I,ppLdEDKEN}Y`4XE_;~*o 6W<qUՇj'ʯ;ՍkT)=w`fWOjwC?Us#dz[gu5_k[F=S{ܓFY>]!|j/lbw\2˥ap XZE ~=}?oS3UIIT";-=6rc'9UFS5=Ւ{ۧa2&OTe,>Q=E(6vTO,.W/,'+-&\FB_.'-ʜN P\J0+`RǟoPθ2Ĝyy,K I"xrH '0 ^5"+4#嘣d%ܟ]>nTLϛb??B(*uQx}H1uU֚fUbǀ<,LA)-8\T9>+P"Rрk6c(9fIwgtc&HwR$e\j/{4kw|"_j$S;\f^:4AQs98_,2;}sK%7w:IIE­7 Ø=y/ Gi  YYJX^`#gq-JgYz[KExSuw!"E_Tx<jvYۘM1MIO 5ׇڡv? S\+f.xCW}s5m^z%'2˸(l(KVIpZ"B:/Ȓ" !!&Q'Z6,K"0aa0U.KKv "iSJlCMt"*؞%).jZoѡ]DST٥ڸKod/p|%~S|z[gsnqC,i)<1s5?,sR:_[S>) H*î! h%\\&`*XT{@7Z&1Zݔ9&UZ8iA ; R8Z'Pii욷@TB(׫N`[«~>A`o\DAT#,ƘԘ VÁnτev[6N[~$XG%Be ˧w_ݽm98 &43h <a'RMcgh3ArppsE{`\VO,f޲_kZ~;`vOj0׭sej}&Nn;/Jh; r:l=S,-~qDfJ=$mL}c궝1uS?Tc?fto^br=|sӋK0#Q级Yvv}nq)$-u7_t9F3gW'sIƬ3\86>XA/ۉcѝx˪cS B0RE5{GDuAbl{ޟȿgg)obsuZuտ^U$ȿ/Wi 6S'.X?&_N|00)?u|.X*S!`6{붟|9o}ouvo]G$D(w}:3Zxq NSΧ 9hd1!D/ ƽ]|lRaoxXWRgOG"2ZNkL2]\yRJkGA"iGezɃG)IT* hVXE` "3ӓoaFSh$()V@DިU+'BF&Oɉ*W H,.OŧOpl8y`dsYηWSQA "r3bCjKMSJ` GH xz^9{^Q֩PU ) IGH-#[}&ov/&|VQrI&&ݻRV9}qz-#mWz{~/"8E]9S/dz?TC2&ŝw 5t7T[ {1Zdح*Vȭr{p[F8fq a/RR`|1h" a qY+B*j]{Q4BC xnT0hIh79: S4tufH BoDƩ*H#6Cu~zA AKQ]rw2" Xp` !$${xݡeMGA,]f;WgIɪ$<%{:x#ꁕ┌`}J4 IA LLEB~G`wx?ʼyu%)rp 혝JmʦHIm,\ W@iT~%AK uO`=)UGK@B7dvTˁlTO:J6{2 Ƹ›4"`F?׬F!s B?4Di̡Ƣ%&B,dl/1zB!EȊwkērXŽaC֚T[#P-!d)` D:9:K)L7G`&+պ1f֦`Bˋi 7Tc-2| s>uڔۄIXGÏۗf>t}|',/xev#>K3OEjic~beqYC_d^=a@Iۉ-;J1T.LQ0W*J/AY@{ix ZhecXԛWpmxҽ9Ӽ]1B.9tgD]F\ą|gw1@Z2siMqMד٥M |;ļ Heh׈v3KYI ]ͫ8؋^^1n2hHUBԇl1Hgq菖NrR<)ByƸb X\{gqz>{zsE"3Ŭ#T@J+u%ali.J\`hl;z,neWoEc@}fPcyH5[sܬU?.yǝg7{x媱7K{ *[aUS!"_n3}WVvl0+ {ѥAG,L3?Z4|>\1/P5;fgY8fSkKKwSI[kO={6U (H4ƹ C&ş͊Ipb%& qL 4VĻ-ՋYirJiܽ2a@tlqZyfſos~)dh dMn|[,?_ӯTVE: COg$$TV!EWTkO./T(YaHNkn/0Tբ(bJ*)krݷ0_ k9ũHoT牉(mD2%+ GbTZTP ЕF]&9 pei C0OЃ]$޵}9;LM(!q0p?M:p(*+LDՕZHnfzMRlfe{2&]ݜ?tToSS,SӳտqScWi",7T1fϺi*3fތWgm@{XmY2 h=t$ƏFo3Ӛ0N A#R0x|9撁L2p؀ ZM].؋%U>"Q^;Vsb.mj1!3ep$- =&B)--H>*EX*>s !Y+=W'c1yM}9"&M*]E x쳋hvѡ~(=gLHrSߝ|HN$Kww~h,|o+dB!pƬ4}ovL-L&4iDK$!1hA oBಮYzawl\P35Qu-VՑ1WQAMK&cuDGB=\Hmlh̅FXS8ہMц1΀23ځC B0Ȋ3a 9- BZP?[@#S thz$ [ddŪǫ `F(ϣ7-xuu1yvOJ(/Ч ˰Fꡫ5JD.+ !T.@sbқ6@a@s1@Nvzy7nr _'䆶-,1tfK&M6#<wrǢb^&X-D敋Ʀ I"+U 5Y9g~a;CN rbdd+k@r/QZh#,71Y\P- s1f'{Qb> Q@\jjbP4!EYRZcC[cTɏC\L$Gx5Ѓjخ*Ɍ&yV9D6aJ Kj2zwʖSM^ft T* 8SHa D&^ssZ[>u=4nI?7l~; N"*x 9j烤0u'2d:Z {==t%s;"ܚD~cs uh1EhwFhe=lÛf~ [mݴE1wetݍv1lkq=vwӨҰ#VN\}nZ\%`_$G[$Y @r&);^]фK SE[;:Y=Y?߭_fu`|[t=@ա}e_B[†VQ8S3&6\U,h:t}YJpuR O2KC?s*^WH<Dz/YaªY̝zc ~)f>ӻ?~ssX؅y|-E]lֆZt1` ~@ 2g"Sa.JW,n(%Y>˲dG(_U%ȷ]Տ.DF /,EtƟ6q=v˰ûCz4_N)q3M̳(.qkcf9>!%kWۨ6lkO?TU%6/F N6h O]mW?yzzmKw]D̾k8;T-^ݰncj0*{kQJ#P A/6#(hY.Hc) 1x.A(xpN' 0o k.N._fluTP_0Fп?np]]6`]T%Ua2.sn>\;_ Z@*gN&ld#zh*_Bz1S?ޜ'~}5]O6F%RI#:{)aGabZZPK G;ٹ^Tw>Wh {8Ї`A.x Zhe#:r*ƃ7`L ־IW/f?0Ad+cu8P~ 贄ؖWݠ#t^%O=L&mdӆqWnsx;{ kV\Nс<&I#SGV!CVYx%@ +G6}͍le|p4i?٫{y_5#ulm.>mK TfT݋Fc}o6ռUd; JR K]gld0crhg1 PA 3G)gBHȇ+Mʔ=^uuIcqV)d$P;Eam:)\4z"g)PL%67|U'q1ZX:0QȠf[a*qPAKcEAF{K{=(Qx.d;fE;)_V085'r-a FSf=i$JT6mϖgR6N[E51F]Y!qG夎g?.1py7tX?|Rkb-?% *>g8`ldE'(ɐT0V[ilm-YaCx7 -ԑ:tLJ*cR] %`j[-B "IWA7Ǧ[2VBhNj]2ZOڭQ]e2'Nsb˩E`O`:Ӧy*9 xf3)k=})XQs9_=҂+a?|>]6/r'J)&خ3w@CQ'ygq^´qȬ5%pEv>3^V@vzeO[oGע/ ,12`4H8$аԖ7#I{O{~ыM= x׎tM|;X%4r"Yi(݂{e:ć LzWg0U6xvj$&@#眨b.DXKT$w9W-܂ 4fdD1.c-א&. Q cg1&NfwnqȚJ#&| c@Zf8 aJ3 9P.)Zҝ$R뉍, oF=iD I1+!p=;ǩ~{j7ixҗh@mV t=p}N+}TW,p+Vjˡ]UsMQP\ _C(yTjYD%" mqzͼehcP$AMH%3sKJƞ 1 ]s%n'{6x:|nF紅7[s2VAiDɶ+c bt1ש]݁vg N5s]z2m.‹4\AB /ƾ؜to&KU"e+%̉Wxw\FϷ1tLէ%cq!Bn1aluױ\"ɕJ}13:E vҸ b/,u-dɑ7ϡ)t0 q_ QA9 AR\^3;j3̗YȾSa EFTژkf @EL[IJq$&S ^B+,zs%}"Y!'N.K<HTheˉ]ߙAz PAb!cT=x2h%6Z4n>Va#"sҗk~T-W{6+}TcJ_F+4Zm-\PseXS'TmX4y5fǃs*1qt[py0ЈW;E;?MVJe3Vұ)q-$sGY*,n*_i{-aXcgo@. mueKD^B4J[=GW? XN׵ochwxAX[8p ֮}2קQ p`ۧ˫۟/㜻 JuzuD;?df}gе5ܳm!.*5]7$f鞎|vd*bշczԯ͡#y;zMDqѺgl'yϖo%T,ixnlZSg^u$|M/ IѻIa|kSJmKDOnm}k7f2pkgEt(a&K (N1l[]f\#B\ZZ_~~Ƀ7ǝTBIަ~U:-8NѴӫmxŖ[m޾^lY-X &J c$Ke}'@<Vs$VG]>޹aF1JYF6COwTMOvM/@u1 h;MHr$sF)i&V77 ~lҀqac5Nއ#SEJ4MEy_J(.(MMb{~Յpvlq}/e6lf[CqvæXZoҀAPC3b$hh]3`5PH/MZNl)0oxyh tASݮZ{ƚA h>ω졚&u[f9&pRV;ƫ=j6''^` K ϐl۟ȥZ}:|!6ïO =”Gí+:4jIm:iNy䪛~0m{F`isk\vVF@և锒jNJA`u IN4^A Nkߊɐ7]-ns}'kw Su7^jpᑣ& +(?>۠f\DSvKD^PURz$z6Dv׏;^r|0k, c0OO r 4;hR7h:t*w]v~NQΖ8FI3@uJ)t/ V!@2huuuEShlk͓lq Vfj-χOMj};0 8 qͅyOWZ\E]]+}:?ja/_7C߿̟n?뺗_\~x?as3w5 ~Ww5 ~7 o4,MJ*j^їxFK57rL u"-b;]AIP[{"Usgت>; WWՀ6F:Ǝ3P8@ ;>BIgLԿ,}ߧz[LQ^OL߭]wzU}7鉉baXH"ux: ʍʶ3+G@338a b֔қ?W~Z?+I9tV/M=}| KvKiV ߮'/>-#h\3z ݦ}zŷ\!D8ZGT6N#~FOYΡգŗ^G.Z NK-tDgD(~b&~XXa0XS;=@V V,Ccp {Dn+^+H+@F,PXR{K%9>L:YL ԻFuF5PSр8f CdSM!:qʨM1ͧpmI!(D <ϾZ72(1׊. $mT3q'좶.x)^f!k+/ Il MaOcג+L4Nl1g~l1"(iyL[̣:Aef !XҨĤnє![8∤J-Jjբ Z"@W+VaїW}(AŌ%|@OjlAitmnNN@,SI9 a:H]c6SG0!W/ڹiϴ>%İ9PD,fTTlŦBFRXg(/$S-> M -1XOuY,fZ"6;Bg"W\r  Ru,!huc ,%9no _LKMߑFD`ؽvGӘ1²1NR0)<@rmv^cfۊ)g>w ͠7XǶp+@J ƺ,N[X{KI3f>Ns|ГƓ|ߗpxxD{@tS}SJ]u׾Hܿx1wleuχv~܏çOW/vcC:M*Vvfv]-MK{Nnhw̨6@[WmX 8#ʞsE6i覍C0nsi{-s+$n9QOBךQI]db~)N-ze)ji|Y?\Tw*ξ${f-qaXγ2;9f쩦wmtXjt 6b\b+3 gPŎ[`h)+T`mFQAڡtPCd[BHp˝wbAm>fެp{Fg:)W˰{$|Y\Y^!g7!/ڪ{ ~Ⱥ%6;.F Z1U UO푧kŸ"*A?oۼ#JyZ.cDbA?} 7ف3pcgftU\Ir[!싯M΁b_ J׾䟃\2B[ WϏc@g8= u4#qgj )g+ r|Cvn`#McZZ\TyG+5 g|5ܛ[R|"Y=0,ݩ+ut3ZVtUoVߞMs׉;G7ۉ 鏋pyu1WhJ ;8nJ&* i9yZi<#}+EXwVGLepCah8 _vFC\)LQٯ˞CsKq ub!DPL`2,=9$ Č;Gx9Pr+,=A`?&%Z(gUyc+Fvo*a40+zٞP]`؆ "3ydSSb'Mtw. }OlCs*{6ǜǵM%p=`?#W8lNkV%TIE:ls* ŖZr ղraK3=:z[yn#Rشn||z>.f3\;~ЍkBw'nrjMHN>{[')`_λ܄ogEbW݉{ӳ7iC9i\l#T>VP(V(׋3LV?|M7iziƨ~cM{D,'|SaK`(_(HR(sx'c~>g"ZҪb\1 A*@sMx,|vv[; a!TYGQb!#3Fñ`=bcbOߗ]/&|hOٳ~F Q?SVm?t[9֖=*Ƿ`{gx ~=PP?l|PKc AiioKcyxvt;: /apG_E?*Yz@,~x>cHxD l1mU"1uփTpxך .aQms!xTV8&tXTGLX>צ$ G()sds&2܈T|f^_ɩc^H~W ٴqf{ݭv֜wQKw2;ݧ!)f1{Κ8Ĩrv7`,Y^ݧ;{ÝGfW+q=l1دmTgēR:Z@ZZ@N%\wIb/듫UEmkǰ>mKrcM8F/Vktn>+Jg:)a֋)=Dv1!mdw>!? 1]fw5D~(_/ښ柕MKRW"GJ>v ׏FwÚCpY}2i'Rh{z27\h R Ⱦ[%-\?n}.Pv)ב[3x=>_DQ\<Y_5pc&bg*zd>ITQTv9Ͷ,fw} [In's )K FDd4MgE|\NT͡ {r=5^#XN߾Ƈ+#7i{u4 :4gczP^ĕERb2s{o˥C'n4:h(`{ƏVW1Bp GponhKW:#{EV5~x1n;6i;0d9xHRF,?10 Zzp&'܆+8G a0 `SB_|@b&AՆ1\\J$zkvE`%sa^@&}#r,V}Jܻm6jSӐ DXJp1%V,r2#)@2OUF9) œEVd@$402GgS0]yΕ̒.J(=t4pSaHx*^1l5NxSLhH6eWT(c\< >7!:xo DyJ&/Ìf:Tz%5ȊEcȰO$c SӍY;Fi1cÚFsq>T<<~߼Y$">|%BYnwh kKW ^r#}x_ jx$L;â6d baQ a9K&8A|ϗR8%jeTMnpi7\y|XF1ƈ)UHY=Y^;,,\PpÎE$gc(D`cIeS;(o6|hy1 i^}T,4( /"2(sF M$S).r[O޷C3Mv!K䥐*pYDr).smE, 'u(5Msd(Tj/l#3Ja'supF>cZ,#ާR*lx_$#[Ë]7ɈZF@<ʼM/ղKڹ$Љor}4CA3$Lb4Y$=eƗ'Ń\&7?5c7#Oxic*y 秏e2=/yJ.Wۛ=,LJ}Vp;ʧlGOTts=S~]פ>?r8]}`dAj,]_RU♟Ln~dy'6BV1cIM9Zĸ7W#߼y !4u,-<` ]\sxR*v!鴚^a[lnc; !\+Aޚutpp腨 Q81Z'QNScsqg6…:M7~i{ѥ ~p9޽,|>,Oh#DF ݦ7nOP:'}!="vpZެ^WQ|DNVt6,c^);kI9N/^=e'>4:]_V}XoI0μ9NhYiʗ%zv&Xb*~.qw) eOvB8`y>&{W+CXRW "qi&% oW6^N[;iT@w4*yIVZ\0~Qm " з"9xkHh Nޮqꇇ2bYd eC42\e |Ⱦ;t$;kWyGQqu/ba{IeV@*=9,%@F2.a 9Z%k adqH aTR3S .)ԏ0>%_F |9LR!qXgc48\X#&g1]1BbQGHtg]Ӌm~l,[4ƍih4yMb8-VE1]e~q;wƁxl9]APÿ r$4I3K^RR$e %EgB2%BL 'ϗ [ri򝤣Q׆nG_* KճZ.ͪjs/j:]ݏyCZ kxilX(`|U@uQ\LT\\*Q`%H"J'$C͔Pɽt2v0)#7]R1Ҩ‘Ԗz.7Bx(0.ҜȆC@ʒr *x>b .R #" {S$5_-u`fs2N74}3 ;Yܟ6J~ԫ46y^ꠒ쨴ԋGR:TTJU58ޓX6jCheSB@ #KJ" Go'Ts$*SjA!\0[q7]iS8lfVv&]0m]g1V<;' H73ݝg%vTAgX+}G0 `B j#FfgZҗ<G7[|Z|S[ `+zo]=I!GO:YWzGުu#P}Q(k8]l-1t.!|ZY L;sm`-gч|$ `pLA:`T W]ÓVQ.Ȋ?ޥH$"Mr3A\0dhՌxSpu@ Jac ӗAc-dt pN;C{~U C)ws*+ҬHh:`Z醗$ƫ|A铀=R!D"y.m L Ƌ5qovͼ푙oLZaC 2J cR8v 6{[6w?,X(H鿤_}[><Na;>z ZmZo't`<(lM ]SU=.vdU*=P51%ҧsG]L5(Jqc/RViΖf6ْkζ>lOk֐[{QTy=@@c6m]O澍8*@L Uf-/ύWKJwoY74 hzo928ZHzΛκ{h"/~fR!Fŷ&u ʬ/M Q.bpcBlp'vJQ71Sg-VgXJKͬv#\yutj9|AzQs̮#|wzXiծp ~nu9e~A,R97ksV:ƶ-V >by:I"R~vmoٕI]7 sݥpU~Zzה?e뛲$+N2%mvDsn]1Hc4n'ݺG&j>$+.2ř8{s_R)}$Ǝc4N|)&[hEOF`/\}s"[G4KbJ蹰dх)@Oʔ6D` ?J{vb`J$bK+WLY}>he:(P|7W{~v!D3t *}?桋LeU)K1S! Wnӥpsb/3X&8K3fc/s&1r֔Qn9N4Fƶw,sw~q-NWmz`Vkx8Ñj]#@ʡ_[O$kʼoi/sQ|co&[KcS 1S^fu|lgk쯻󓽮NVʒCQ5.Re9]iPdG&G$b6*#'BCA`|J[{4wh{/,T( 6<Ku<OQ~OVu(nR9ڇxv#O[i[7Q ^8PYԢ{sxlhP_&֔7`l kr~qWh4q܁QE9&5E!Vhjd :qSu)M!YKݱXhC"1KѢ8)J1&e!6T2@=1FJ!PSFA)5QJy RZ/ 3SyaP幧^±+XѬHPLTX!ׅpUǜpFQUlFksbA.R:ե1@lGC,uo|%>>^sYJ[Ƕle[`_p˷=rEP!QkvWmnܭ>͝ 3b|E&?UP*rÑ?ZDU#ۿܽѩ#Z6)U>>94V]M6`>+v}R 8xW*s .Oz)w/,6R$y'ɣ7L7p]__DnD}sLӪ)ДǙ*d J 5JNDT Tc^?nV:Z;_7V;[8+:Ro߼lׄ~h B/Y)`SpGYBpɧ9>GL{L%FnCYM>duۡ{rݐ[Z._Vk{͒պ~zt.F/Y8XYZ&鋍9tQ<@`83 ]1p#*ύΫ_g " c;, R8|Q}YwQEOb(_=PBko?bZ,g-{`v#l=kM8 K ηEeN*:q`ꬹ.YsAJ̮c4z:kw(kh'pzOAPu EtD;ՙnC[hz{ Y| [W RDM)wK}Ӈn[hEdLɩ`4:K;z~\ot RPCytR:fȪ'M:YЊYΧUa[ұD\Q=5[:1ES29 Xr+.YRj4"K$_٫SS#41jkI{ !%a*E{5zիwQCր7Nтg`u&ڠfm_KQ)9_PkBYTJbѡ2/-tN#sS_AH"AT՚F<*K_ <М[屯Y_Q Uε@M"N\+^qzux"5ErrA(So6Ra/)bU3əɉ@Q0 +g<qEXέh1IכA`:;, XNXrg4o! >x cZf4*zK(h2PMxN\nUcb XwҞy{f K?۳4hU)}3=ITK<>)ݞPVe j~ӸnwjCL;-%z@ A0;th ^iߕ:JVSpFmV]xX͙hl: NAH:/wy^:1S70_Ƕp,,SB1tXV6МYb - `1˰Ncʯ7e]V×^==fe$مVNSO5 >tAPJ" Wx%ks|"6ٳ<䧼t^qF ]mo8+F\8a)QߤcN`K(KD=t;b=bH>P4?[{q#ɳ{ۦꏵgS4?{oG.zu1dt.6F1x>k-Q:c|~*HݺZOWb0rq5,]3=ʬ\r-D W,6̤1=΁="1t [4#'hħ==&LHǁ::\_< }k$~VDmDpm'`¤(!JeGn$j>d1H5Q|CT @4}((o-k/?A R$&Q8 [ô$AHWJ)TAMeo-kzPɧDe haٷZUb97ud-P<̟t",]-I=ݑZdZ`\XQaX5dcu15ш,N ;C( 8{(!ʉ揠!ğS>,sWWy{}pqFn@&>ͦs&ɾ3 w ;u^?XWlۆ?@!:#{7p2huȻ<IuxR"JTT\ɽ&_z4MdZ9(N8FX oB̅ 1gA`1Hf)JtJ`7ߴny벞nK!WI$:_՜``TcpOΎ'ONHt羨8yRr:s_ 0þoȕjSa#%h )I hC5$*hC$:t-B'mVfaarA'$Dqb:dJRްn1$J./1%G"[ A~|wv `]߱}Wo(׾Qߝ~ؗccW6sh>}~[onoޖvd9ra0A ڦf OiόzYLq ȀqC{rϧͨ(#(AJk҃BL76$SFƖ ' ȸLh*3-T3҆#aSiCQG̍u$ #w˕r TG/C%!4 y} rPBY`wP~761wr={G1*<YuA}ttJR? Ɣl0Ƃ4Ad.+ &S(8gcROwVPbKLհR=A%xg4RzJ~hh9&TⲴ_fǥpvyVh2=[BIE5{s3d=WjCUSj@v:HJsɽFO$8^=̮^˲گ ׭tqϙՐa0T1|R6r@8x_|֑-{fE>%<4A4}ǯ3]A_tWt\٣xsReFឍ:eQ NʻRBw\のITAQeוyRS!:aQ@c$V1d}E JAlz-}tAMe"`AgKqs nu{"h̬zp"yhw){٭R[sI\ד|0>t] hdolFef+3nqj pf$hZV8D6qZR_wkD8-MUN#^vUlz"O#8^kZ-|En Խuz}Թ`kqiS)Pwkհk5PBVsk" H؈z!=$1P Ŵ$ EgnLyq }%7:._sQ_Gy!}BwT@=>t1a:G`ܿ'Gt|yF>I:R*Kn ':mt8q >\w K^c+NZ7-ꇴdE^WQXKZ:W>v^;t\PPT-)z}NTBt655ȂIC$[:No cmhHH(fk@k ulPx-er 6}^*;/7 揅yf5Vɦ;sX@9min&_ '0Lw@a 5TcFXwmco&qyc6vS?*q;G@vRΏ>JX9E"^|U?([οHM?7kk]G2qlɫ>EOy"FxL/ &bD ;;s7SnV1$r-qʶ=D%[L=Jh=+uv$D^މ^%k$ݴLxcF@KјYM"tz[gNy(^0U6' dAw4~oHSHwH?A{NtH<fН&za]FVICSAh*;X7;y1'A>DgR]X]v&GqCpΪ l, 1Cz8_,DC>"R*r! z clpJ t ƄK-C'"ZsH `m]fs46`' ¥ZC=X Blnnn 7?gyuRQ}YĎ\3<Yfq r6\_79Rn~ '+ Z*?#R^+|ǘ~|IE}ߧ/srt}gD?]= ÿϾ[;n-%J*I 0È25LXSʸĔJ v69cPZ v~j .z/eQݽ% y&ȦrBZN7o*,5UpRR$"w/PEwdQ$GҐYcl[4|ŧmN҄9WRsfˢkz~9#fUT#CP;g#b6z)1ٚsehUji旗EKt!Qp$ a,4ҡIΣk⥛kqxGHrJkiD)۫(T(?ъӽRH{u};{WsJ~lzb &n3?nA,gh,BJݲXr$G-nY,8aR,jNV ؤ;Y,kCJﳇIUa7/`В%#0H`W FJLbJT< ( +;"iaFoRW_o??GpoKٷwJ#w"Mo .3 G҉5]lFXږoir^Y1l]||oCHnOSְD:YR2w#O)äȰSva=e2?z (a/ %Ӏ+1}2NǏEPpfGu9+͙Qյb@"K!g| {.'@hjʖje]moTn]Կ%~F 6ț&ph v_VTke&H݉tB{=3n8V!hNw*4 YOD(B*ݹߴA_}3j 0ڎ7|,cv1LݱluHpʼnM8BH$vǚkf-h*;F55!M wd|/Dm<'*%O?K?̺H g/ۏbD!p#+xo&GXpr CjG|9CEUJ䴤掿҂m' r;Υ#Ͳ5t2)R<r<|,#{띉&OS`l\>C7jhͷNwgo^L\U 5e".iq 靖 D8O$85s&IP D$ 4TJܒ(hQ |=z?Ad #Yx>plϊ Mu cC=$)`vv<#0,{ٟS1a&l{pc`2LbfQt<8Ϊ RܪLqq06޳V809ӿ:xNwj~j|ؿ:wo`#9—׃ ;);9@%_xML҇`I^yq6:^g3]: _?Xgh z%<^Ipa-tO]ȻK ~<t3%2g|}v)E/gf͗ߞo&;;H:zhV=u-捠7I[U:Ave /HMlCfI93G|wOhWpiF1%2/5ҝ%FTޚ L}ETuߞZso̿?'0z~73e-?wuS ߦYp \w9 q9f=KLj%!OZ neJ\D"̾Wj:*$ fmzf'RRr?/ol|.\Mp惿۪]} ^j'ȶ@d˨}t+(8+SGi# 9bp=+7떊l1ӧO9];ankJw )A"ǹf!"S`,[#_8"(ZJBqFx&iu- l\VW;l~ֶcii0-[4[/!)߼=zI;TpA g_׳N Ƅ(`A{ADi/VJy_i}p!'7^L@\<_p:6sM J^bvFّ4#qJS \*ۯuXh.I[hOPyKst׾T2Xj@eHiȪYnG֐I.ǠJm\?W qQg]l BjUɥfn tryp?qWI.כ:&|G:9@+?Z;M}%r0|,-_v|-;*˳{(.Zisb'(33g`2/\'^0x=WfdԳ=6ίXLL3[y.7Xʵ/-k)mͶ`ݕ?T gͣ0h0ڞ%ФY<}=K#J3&5/Չ9=C8e,3zûY͚%tG 'ʰ>)K˟>iD18PLֽ N(l fOGh+:XGeb5"/.fBaFٴB-/4ZH;2l:I/g_t L{f63[+V"oD%2A0mid0Ꮐ2'|&uI=Q&twneŃp~jCtg2AfK+m&h{ʍHve!¢W F:~.CF ң|?gkVO72YeH=̅)M48O Re>Ǔah.QH<ᠣy9ܶcH+C +p4>ݑ_q>8bH: Cj$TEx*8D2A@>e2b8| `0R=1òZa[sr5[c`"{+͓w8*@pwJ;;&m4X2b?0JE*b!a'KH,EdIWȮ;ζƯ,y;M _h$;ԗ,`T6^ޤ$/\W Kg3,;7h4) 8lT\6* 9y/_kޒ:Y׳13ItwDK[Y8 _̑ù{sI#qb.(:t/ۢvo0!!ӈ#=☈@H0aC,}AYO W0DJtH"1GaA( "qx!|@9ҺrL+p;H -W,"MdRp; u71Kc!ΥH/[z#Ĩy" '>fP8#V&As N3}EhR۽e=aoR?i pKx(Z. :v BVXFBx&˜Siμ YFPMMFBDFrp ))B]Mx3iYAunkܪh~!AJ%OX#(5 7`)%>6#M+5Tpڄ ^ Fji>|w&@Ȯ{[=/7+~%ȴIY,./+α/{0UV>;~dH陞!izb J?-*}'ړLж]ĩ2R*iReUYS +JƁ9{I4ל\7n&/mFd,}eFXf\j*@jYREBV9 SP)RͬGNe*~1bTvHx}CI`ZQs& 4?ZB# 7Nw-|GƑ͞檚6?Z{_*]x|| =|'xN`>FJ!&6dcp@F03<`b8!B}40$;:2wd[i)E[ωbOc pw~.#rj K|I}K i5Lp8~jD8aڒ6.p;moӓkq?di_$ tZ9HJCf9A!ǡBL s>i9p'o.8@)]e0/a8ipneJA 9тK}Jw wuSDut?N>X Hp=ĕ=bdo Wej՜1=#b |pb8W:LN9xB/a!Dl `z׻AoF+ӉFw/:'n<M4ʦ4.>n6;R^đ͞qHM[|aWn-&Aafa/l~+=d6G /Hĵ>CGWl^> Y/qY"jt/'>QpCuffh8Xۺ@#o~Pޮ]`uw9l/>\}lzռ{S7r=Yٻ7r{_>_km6{<Ͽo]SPݣ96{ YuxX~ ҷ\{2HW5 _ zqs(6ঃDṔzN?՗;Tg}= W"Z{C5G8[)Za{[r _WP:ӶRiT})`Vz6J0}R9Գ[$+57T+:䋰RN~VZSwfaܡٗJRΗ ZF?Deju㟺*ϼTA"A7~7,ާ[ۍQilJ @S4`YأC`@P%YeAbVفJ3:&vOę׉f ? Ț熃V"Ud,zXA LhTVT:V\s]dKZ $4W12 )C/t"4|CC  #/5u|FMK|Z(hyiF6|}qj&3<| @>܍y9&hU#k9rɂLI4\y 6@ԥ{j!, Q S+($)rsP*Xal9t*FBX떽N:uS-, ̊ JUJYŐ>Xenug)gJHѳ֪0aVkQR"XX2t \$KEdw҅!N)P$t Adb0V=Ez k* Qj㘳LjJB 2ֵ(vAfQIҬPbHj6n{4Zͬν5{Wis6_X!p0(9ͽ=}F״>b M(aiy%q@&=YG-޽/%c'd;4U3՘8eg{~jax>,AujdlyFƬ+V|)%V'*Mm#l|mnjR>a?g? 5r5xWOͭS[/\műWpj"FRy) #b Hq/vS& h`=0% ./pi={ dhOjM쟗?~]kFEw%wWy*{J̮X-Gg?:YM޼ YL,CfX L& 12S"4+RQ5˿;n\o~fxd56 6UupO sJ,CGveTVc%\YLxV$LkX%̭⹛Ҁ¬rs2]@<FXc30n9g݇hm < 7I<\gI2ϕٻǥLs_|5haDӱA"^vf"GNGgKW7|ڿw^4y/osb6k=/>|rOW?^W}2BsbIs=yڶ\m۟?L4:euoӓq?$A5hP4M)Xм .e35Dx,6AQ3&_nnh?{Wmzݛ(ai1k4pc}36h4IP e=TbKOdŒ҄/dӷ ;! c;'kW~zqU_]1Zy(ךxU ߟȺlWu9tol0hfp/P&S4-]~S2

6nT!+bZpD#Q ٢xR]4s$/9thC{Y簞:3CxMb<7!gqPZ 9՗mvVzV xڰpYVظJI*j3-T&s3JΪ4Ym-T4;T{mp4n̹VM$uq<%=44l@?+l'mr?Y}rPzw|lnUq[ilMck95%ݲX:5{bH5Nt21 {TBTJ&v"<-a ɂhm @eHƅI?CWG }ZA=VmWK?@% 6pHZ v9G.k~ڐ$_F|5E~\)X{s8ۭ%>d4$*gb*SEL由` 4|.(g'f NX&SS&VBM(kJ mbFJ!}TFxg86+_`?w96UfIٚRL;svrRm v&G R#}NM S>m;n):0iP6I?sfF(o8vR!)([86w@N(ry %~*$Ǐ -z㭬 u%J}Vd QR|Sg&W]*´d8,I[΁0;N=4LB䚋e)S?3Ҟ?>£&B{CbQ#gTJ 1d 2,^&lG" uWQc0D4w1́K; p LEQ}Cs.PtZCZ Қj%-紭HG&Y'q.'"| z)/A]vgO|nBuc$sԁ^ze/n:p= C)Fn; HܤA&lҍ% 83r.'أO/S!FJLaf"ʒ\4"Q$ vE<{ iF4C$Y.sŠ sΩB#\\R*M)Tƪ)D XGV_E4fY%4/c5L0]1BRSQ+T0,2yUfE. K%debYdT.J#zhx5F3޽D 8iO#/c5B$N6< (2WyJP Ke1sVjid S#FJ9PaTCaIzۯk*"2BOrwF2Ӊkk ;}DR͂8` *DoUnޗZczg~+ݞB3\IF,+o밎.;vW| 0mvއ͒y&2< as75E)j%h89>s%!q Kì3ɁL$qzH͚!Y1#>|6;ϫ:R7H/~Oya?s?e|4~Pm^n~Etfۚ/Eo;ۻ?'6,r'ޞ09\:FGD|"EK#"jSZc7x'6~jpWuw>q|3D[v'p:DN`ZO@!պf/cWwuɵ&`i =xUXKipDž3I4x|S K@hrkFGɥmCe"DDjj%5g(=_){ C}jQzx(I)FA )}}$"ZJqTHBJunflF)KE}K셔R' j,}Rg=˯)yjAE/}RJ2'C郔Li5p_UK)fyR( )Jf*1YUO- mp"w4rGBY(t&aT? iCUJ ^{/UIh3l~G̢ՋRi?|ĜE6wFy卦 )'r T#(!KIAsm~g$$( m":,'CNS}|$g#kl EfQz٘7|ecU^3$mO Q-[ ,ggVPz=N~unM>0){MS鱽h6"ְ>ׄnfe8gͥ$.g ;R8&ez2QAXOa=DxgZ-F֑]l[ ф+x yM* R#5 rkœGQ9'ˋ?dz- +ae]aѝ0TQ~%g*|>|7@8um~fr031lJeZ?m0yz (G惨OȎr9p"Tz=Sm3{ea厙Ġt?bg$'Ie]*kZY-_w.**LK vn8 ֤VpϢv#bOPmkm$bFE[ſr_GQ#ڷJ{A{LΥG#9CtBD\ Ce 2c@jȨ`!!R,Q 9fG HX.Xbp[mON]-MB,̏&l(GZQS)ȹm7 !χjv%}wzB( VzXœd~ӱ>K"TƏݺ}qwG)Y.owEk3p%QT|砤m~Z'O=.ɺ\:?VOJ>Kkj6%U.d^:,);7Kb CdJ#H-9C=GhAHqOT3תYSJ O Rm&ng£.X>u^[[ w]qfln=k݂ `#da#/xل]qO`yܻ\q(*|-ۏ[[˛¡t5rQoխl1"6"DHÎ̌1"j2`xĜ-SkEwpxn- ΔZ8HgC%C Vd?H3˭Q/9.ʱRi,H2,a꜠:nG-CDH_+Hסt;Ⱥ3ij/C"6r]3z׷VCtC 2}3{A ChzVj/-|ɿoR1yc"Άo~fnՆ?O&$<$ŸBCo~P Sf0?7ɞO 1THP?CHHf04Δ:0[TQVm3b 7/zeo~ΞtCۃs&2Eu[U;ƯKꇦuD]G/ q2*fv{WURlNCFv Ud7Oz~|7)V]7 O7݋#4jEށc*ݹW˰W י#VZJ!I L'+,9,lxdEtahK CGxL$#O#YőV_Tرq_N NdSв!Dx^6Z8¬zr0Dž$u0 N3PYK<X&H14`XjBDC1{ "&ǍK<ؠ"yp֤Tq ZYR|D2AN2J/1R(cҸ8/8bFc\"Jp[|]J0éCA%/\,!&}W!CDEu2džsJ'!JnM<嚏"nBBw(C Lr+yMJk"w̫!'a/#ys T}w^@$Xuc#P~,v'χ={!q\Րgalƀ2 & , l"V44jZO-֒3YݘD30|#(Gx '%e3њAU E3dNX¨]I RR($~2 ժh]C=9zJRkôS,(cʅrJ5hzC㒔}C<b FJSQPX;t/=cnMV"Wr:ΨCj svIBC4nH@vu:$fQcFmďtHLE^iiZئqC*P:&'Ǖ`r$qɷR_WDTP!{dIO?缰2ŔyԦ<娺.U?ukMmZs͎C.>@λ7+\\-/=dSN?4ۧGw`4>G9ׯ!IUL`ej sbEL!LBڨXۜSǂԓ Z0x=wBh3pP4$z(w]ÛgMnh(Y=ڬ]yBDo*:gOrS&l@S2au :TxbΞze]OUH({ 6RcP}#h6&_ j%cX!BɨȒ= Ό[\ ZRƍsQoU#&0<IJcvVH+ɨFzt2*RϓGPJt6L;z~V +60!BP%zp͓SWWT2Y: >#]:@f˛jNGRL;NtTaJ.흝ݹt l 51^QE&qШ-źr9TXMϥbp$d>_!hH ؇)2P=BX&X0 z g腹l.PNa*"ZaTД0*0ø2D)muXY,ߖ+vJ/` qN4zahh.8@rep~2_]R-W}")Z&Kioz@x0P 5(V{=KmDDs5C4MlZ]Bٷc֟nr|*%$@4@b #^*M\2S%o'3.#%!DR@*"A*ଶJ߰х Um-??v6DN&X3(b j5wu*Ds#vuPȦ/T{;uDdS-MzQ%I>AG Z:(g4iŕWu1(NžK̤;L'x%͒j3YOc,>jxclT@U.HTzU4C])ǪizփrkT ).nF޵#[Qbp9` a}څgXI~$߯ؒ-9햻 (V_Ud*Ȁ7X:繈) S< 6 Q=@4<(!Yjh5!2>CiN~{S$= Ʀ$[%$/a˒mO%7K0|X=[+{6LXe;Mf$3zfɿ,8גE4ƒ{OޕBY\t29[͟Q(sPPTG.Ad Er6Ya|7sWh#6qN `i%(X\▥",0$d,j`u"h9H$Y}ݍR$E^IC{t=#)cJ=+Od$y#J'|1xG{G'D#FYF1e !E2S0Ie)_2W:aIگU  q՝,kM3<Q?e2.xbL8tU6D<2QNtJjEÝc?uF #ޣ=),{ ܳ΢0U`nbJJp޸2}[ݭ7Ӹ:c#@AcHÐkG69Mվw| 4[KH 3VHKDȊGösvqRY<=-ɹ5$Sj)B!hv[ԃ`@:}2!rF^#'g9' W YQ|LTv@,>Sۈj~D!E-}"wQŧ O Pr)N;.Fl7N:M1CGdm :|=lfqD×BYƁdztCWxƕ9$OWBI[FD(*F~[RzR!dLNz fMrn𜓀85RY{#JBsL; ąU V3҆ĹhBdMbEi%TT "ϯi!*+co=^܌,uG眶j4@ʹ8 s[#}8a\2oYE qpdWl1uz &#QuJi0eiEZ5l9eWjF1% wpB}{k5#Crv<9yioWw?3z Шe$$>H&U"E#O= Fuť yB f2+Iqyc[ Jz~@)z?{!2^fsMALbqiZ(0"˜-än??@}ˋO뭗53 %S52.Y{6q |=@' 8eSH!L=$(-$+FH k{mH[Lv4H%@$LG,Koet2>i!L]h$w -CSw֤Lܕ2&}S&vX5_Eg]`{CqMbY +MȾYγK6tpQ KJ:0uf$5 +m izv5WӶ}1cZ>3n-Th1\WV !h#.hdi ɦZ \}o?Bוּ>@% rysȠՕ٧rFE 9+>Lf&!#őc|W$by9rͩ)ٚKٙ"]@sJ }WiXj@8g?@KQMje3$}o%ڦWOHnbXH>rGppIPrWG~]͒l]倃Pav6 { TG{d@Ju<Ȟ{2|ARc}}1H|y%,}<5smkV):S633҄XĘ{Mܕ2`}SbKusdQ3$$ӵ~TȅÝC_MW7y߀ sxuQ[m.zdkNP|`!9 ,\ɚJof~f =%=/k?WźWQbi3LAaE4"*-nwөe-pVnMhn۷vl21$knr!nmԛ՛Z[ogW ;Կt5hߥ~[avrU 6kmқwUiU6ry ~l >(mEj6x1J#,Ԣ^5>wINd}OUZEOw i<͆g!z+8&e/&$o$ިڑV6n|HI21r=ek![f ;ӗ Պ9Q|Bd?^ۋVH~ NPX-Z?˹9>40lwZi_~h~C ֙=c [^.7n@"h$~5'ן˫IYb6o\7gHWS=}k!d^dm6ODR PccDF'8=ɷ˔_.$p[ 8 6IKdk5=O^==}߄>}K&~ܙׁFwHWㆆ( Z+TVAڥd2,;eRu)+r9y[%(SO@èw%TZ:,:Y9@ $&G̹I̧R1cP p8]#l_(߽.[%a֥w ޸Ő=HxTEd/-Z>s6..4DGC_Q]FOaZd sgIFZq:q'fgXkBW}0-Vr[,d ~%kB&?3grS@)͖86hZaR0jG!ΰ j!x%jZ?7]aF+̞8 ^( ҠC%}!hcٕFҺD0~S.Jc6F(r heSůJ]`s!G݌ukcB#lr.G1DKPfWXp\0酂${o |5kf?$ lEOɗQhXmtrЬX, /"#z/Euk'hfҗȤ$ $,u#j ^ d-=~0ѧ`m4]2)$F_|UE)9 wD-<#*]}aWw0p|)PFk*1|fVkAE?{oe_"lV/x~9g;YݗdפT,XE % NU~Ziϑ3$î/>…XkQ0[K /iJӯy03wߊI+F04שp%QTj<|< y.L')+F/9Mn K_[sb {ٻ6rdKGH$x;/>Bh^YY%-[ݲ|,Kl*ɣ#C/{y+|e9齊Xz:1Y3IY1]:4dyc)T`m^+{CB>gR{_!ιi5}J7ô 0*HI`? Ž5}c\U+iuO^%N+}o3c{ ~IpRNP 1GF`d>̕m;m8 E{7< &ZTLvJdR91 z(br!__s롹KaUYP&K'_W)#JCҽJ[ uM6R@^IJ[E~,1M?X=\mhl}{/R$AO,NIKwnKIka%v5{Gރ%ib2*'Ci^qT yzm8hd%›].SX>랢[=Db82q'W]O~?1dZ/zz~\BGeꖗ9fBK@^iJۣlor9Tx9ȷra aCKv+;>4fS Y 55U8iw 5vINT?lGjxurOJxN~♿`*lmjAf?Jg=irxo>ف[XeV"s?u.}JhmQ$$]jua(q"'qn L=}*?Fo#j6%£c5h }\[onɡZM{`ⵄ-Kww:ץK$O=]Tv\=yA+Z1;r֬A#RZV\`Y+->1j_ێ]Kl>kUIT?~IlopgZ߹W^bv~'G{יWi:}fzʋvOf/ Yqg~'HԬ/f3 Js_lgJ`Fg,f^P'l@gѧ]*q=q,Kh$ͷEk^0;'9.?W;ʂvYHMz4rqS֓p$S]ݣh7!vJ&m#]S'ݺj.]dJRNvv{j5uiN_ɲL=iц\.<3̾#Zbe{gXqG~+fC@ƺ%%Wa]E;hkÈ Do%ﮀPJ (z U꨸g%BІ md_L[bV ^=KQaY Kw*qHTgKʧ(Gh-]bh>VKHU޸ V0d=Gqx%גKsd^*yI1xNE Сp ƖԒH2-_c29DPj4TS̕:C7EONYQM4>WmV1r`Q*X&0I5gmW i|nFYe/_dgL#L(eo2QR2ua w9]E3^ACvQ(p\ǒJ+I ]Υ0:iI gTPUܿ(uCz$@l5JWR`|}k-K U1g˸ A)uR^  Y^zXWڂ)EdS23 ,GVUA0Ij4QV`FBZ2B\ mJ{-Hwd?ܞӵpz^?>R',@9.wLھW~*сN7k!&G,pq/iW:GDG%bV49 DA6e|H&5k9ڡ\0&V:-U%|S:x)B`N6:\.L% u*z\iD:SPq$ix'*`tΥX *]q% xdNK'p+DN#\ ]IEsoQ+cN١=coFKȜ)`t?J4%F\:c#.۝u_TDTF ;ܚI-,t)Xp[[Kg0 ijBtf%fiIUڞz蜆fgp=`6GH<JR%eU YBHJ;i? 85Y!v.ʤ0;NQs2%宬Tq;FJ͡Y=uZYɆ䈓 YZ`2x/Yc&f2/}*f ' >s氿zؕpد~O"?8&&dkldǸ)6k 6V˭ ~k@`YXQW%G3ˁ3b~#4U z;u%S`2j`^64Ŗ; [@* {7TCІ1Vp$ذ +]!-<-2*K#J>IH\Z(6K ƘʍB )+:::-lD:TtKU͕C?6vxؔ9]~ǫJZȝ  ?(*n49:}FrKV? { lvɑeIV|g~[57;z@ r7{lv(wƃI'P;$4Vz<""uw<P7{zEU#Ky{7P%Cqivv<8}@&bk@`Jlclth  8B--mSM{D͆V=εT !V^-Ye.dƇr۽xӨ]$-[^A;Hrl7f;Ym޽NuчԽ[^7^ֹzt/nQe?qVU[>ccvnFkU;NL_wގvmtB"ל <ĀVQ>ВnΓYڱfYwx֖;fbwl?=Cm'8 vΟ|_-ɺҠIhӟ6/vL'CB^v)g`-;MkҠt}F֝1ִ[حڭ y"EƝ8h7K֑6Ǝhe=hpۭs[EtsrEv>zax{h8!.{tiH 3d7RN 4k!q WIx {2ԌuFz CJ`:xل%{hN1#dbhCZ0#eBJ"_u4iK Uj@=) _Ti+V*J˄Ϧ*򤹶n9uqwVg {9hA ,ۆxH@ ja47G[jIUǹag̒{A?D%t%xAx!$Hf8]]u+1+Z# [ KrVEc(݆n㽣) gw&;9|K4f, PqfeNCpH] m?WKo97V\>hc .k\qI?.5xkL|M|eSI%;w_r 0c{e{(٦(Ji+-ll4DžQjeD6_G ܞc\bZ 'Fxm9GȫSBR:؞=TL}|/Kd.Cenkq$F@.hރgCX 57+\1EW+Qn'_ U*s[Ά_=YeŗX~WNKs_oAtucǬۛLzn&l|>m]3W>i߯O׸o|l)Ksvp޾yXeߺY[Yl*>{wgWT7F?<ϕri㽯ֳ[,PE,=UTwO hahU>h^Bl j$Uڇ7\Bpa;ka4ԱPoqky'4X%\q2 0,XH11$7X\d E3ee kV>}R3i$Bn(-˹r!j$hbU{mNx}tӽY⯿^^-OR ݐZJ-x&[_fMjZt<)ZH,B~F <@ CTV~u𥉉ol߹?UsKE]MTaͪU8@I)xA\.ũƾYQ`ƍChiݲUً? Mˬ7~ԋz`&^'?86x9BX\lb!li7\RK/|Aٽsyog]ݹΛ?Wau]? <߾֗zc3;/aϚ}n"}yX# '%5ݔM[⿒nV4q*XI5N&p'$Hܥ|69(k8Þ'H%-5؜, "l3ܣ֬'FHzU^7T8V0_eUnvVW)æ5{švVbk' QI|_vFL1$tqqV)6xc\:K( AW@ZOӣ})hE6 Ew)uI& CFH_`IߋIjpz 6[@uS}R_4:mTN+~dVz@-M[OVzVJp콒.g+'ESj$:YQ[)%aVJIkdQVZpH^"#f!\6U5T@JlX^X[B5RW3>eQNnQnn-F (yN"iR|mLKؗ7zєUZJ{r&w`"e#.U#m"ERzŵl>vʳ?|-7 b˚# U;!@RH H$YiDH[)swu,DHPN'_)ԟ*µR')|ܑ"5ʵL&o6M3D.~[@ҁzJ jэq3~!.ɡ. Zs,~!v_a_~!gRKbL9+Ϥ1&# Xlَg+rKO}UoF篖}&,PLT\F .-g-s)mh^B)6(OsŚX!aˈR핳y"̦C a8RT|.JbҹZ)B AP i2' }%!b3ݪ .sB\ߞ?/U*qvԫRӝri|sZo!+2KS}>IݻsA_=>s,5.?GB2'r>gԦӞ짢ykra[<0aҞ125~Iyt% lvX:|CF8omvo1|S9HPF3T-T~qrŚ/lZ]}EE!#vwxF#.E3{3N,׍bSHxQXpԇA*D[(`J@T'i]'i*O>TR_ Dvw A ++k wZl#H=.HL..LjuNN-Jځ{l3f]lI8+r(gŊ>Sbey_؋M=4wG|ql߉y!@"I{˂hP$.؆cՎ0LJCGGC)k|} JTy25Rkb3*f"+;ࢗ(5EOd;X+ yARNKQKߑ"v8;["t D}jC]Ǡ4W'&~8($~>+p4WGa"C5*-q 0zN %HKoY5,ذ@gۖ&^V1 } `k )!-N D2(׾+wƪˁ~wCYCŞ3E U #rǚ& L%ѥ, &Z* Q߀NQm=p6]sT8Oʁ\B87xL#x$n8$! 䃹WWbpG!ꖣV=/|<ϷSS=W1lJPr:aА qG X"x*ق`8-}ݹ8-N.:Fq }VCξS"uؕ8R(OK!6j} i{y9]''muZ)M)B9L /1$yU) <=r$B^&Tg/p"=n+eELxLtM ƓBRHm.1ˌ}ܾEyY*=uvXz)L$y-ᧂd?Ug$~,5$݇ 58-E aS' B˞Li;{Ul9I(@^.\rާ;} */_}5Y߭ o @*9oy{|nˇ$t>"=t4P ߰1Y<7`Y)3$3$Cbr$yOe'[ۨ19|v!_fR E}|{J?@ ssRqpBq2_y!o6[L! $QǼPڡ$mHP6 aS )LEy$WSbEn9XU @yiPPDc\b L 9GDuIy)p##lQ RRs4Kk %]( %rD@^Fjb%Y,ץA3d΄aHEɭPLu4eDgAB#piQvyIyV\`rϔ@֖s!7 \f;N&wǏ+ B%;sk9騽ܩ;ZI4v7Rܨz:je췔k)5K%8Pv( {n2Smh4%=Ma!(<%_ tfK4aq#%=(GuR^TlW wgogC,ܽ t@z Ǿ f8# }GT%C%Y<:uO\]$?3Lĝ3 S]qR=1LǢSֽ\in=ş%8s醬$c2D>)t@x[ [!(;'Gxult]7~Dx=:0sʴ|*Pg %0~_f.Ĩd ,ђJAU!s5 ;wOӀJ \Sq(=_UGM맴;@ʹx^ŷʹKaoarٌfd@@VU}:V ]Hl4)f@y, wR5<ݏFB/'TɼpU3OÒ6:`+cHH$8Hyh)KC|R'mLq=Z'irO ;dש/3 4)&I/VO #VO=%wjun:Қ~^)Q&dd|Xu+cJKf|=$WL2z^Sϑ2zi֕ӷ.ow=|aqO^2HnK^/KO\vc(I<@q'N5dS?Tݏ[Pʝ {k3@gz@q;i:/4 ܯK/yNr7$2d;1FԳJ/4 皽 n0*&{RgLh;l^D~{6rŏx Rzg^ H'J/-/rɝo^xDoՑK~K \}xFYY޾&/.w bll u\%bR Ɗ2&bj6YAVA݅ƾhT!/|[ ިӏҏzz8~I0^N_ߣРH[*E\RM2hL]eqTpNٺR(**zf~ МV=n/׏FwG_ӯF*1$vi9GM?V9V{FӞOkKQ S 2{[cHR$5~ j Si֕gow=|!J< {St)ח)ۧO.1Uʃ;$IŁOVYT|n2jw^rjMӄdzz8ʼna||-謧[߱SN=>?꿞,G,ߞ;SҨRFl4պљGeR?l.[?19;sZ2,ӒavEqJ"YEe_zβ" #Qpns{a).כˈvb:H,bmVBv*HBd}*}/FO/)(]h,*zIQ% RPe@)Jr%{+!Me5eB|5\Hj "|ib{bb6^_o<Q%[r7db77R|yd#J_Ih) UjA:(n1єhN|ޜWΚmofuoq{)H3 @[2LtUf a0YTlBYLHhb, |\`2ڝ)1[ ŝaLO);FdJI%,YI3:"Ppn6~]qU6Zozi" 4{ IĐ%"fx FS(u"ͷOPE-IF;& ƒ}: }2*2컢|!5dN> XVYWXdl8iȼb}lAY7z׷W׼5mYj ^"`Nbxi:iΔB.U*iMū%:x$lH^DY UqqAd@cWsQڎ|1Ę ,QAAdv6Ҥbb N vmX]t1Zh:eEY%096 *+CQBN*].VZeNl_ֶa@DX5(Z |,H+8,`엱z*Ѡ ȓxȤw)bXϣic48<;7c> 8<wS9JҸf-4 ] (*[ ! 2C/ڣTDJ?1t"Wlbͫo\)Zgz Qxy3uWA\cm' c "^XX ǁMibo&674շ{mU)"P1qT!m68x҅6K?QuJ,3 vy Z8ͦh ?L' ۣaj:] u 4Ԧy[$9[X xxZۏ6 ' ْeV20+e0@_fV^mi,Z?VsAԏHS缣*2J#e SboɸGIڣWi(H"xme)xc%ۅm}]o~ Τ&D:.f)bHXl!8"Y12-zK>=VsgYd7`R bj'Cj*Ģ[.ʪA*ETTpOq'\D5ZB1Ifd`?JBʿЕœV/# пX3'ۧFV""/>I۶" KncvB3v\S+B{N^#!J(/M=xGP:eOc1,4سS߮]y_! _Vdl 2u"𳯟oWC A?^4VʏJGe]] f`7h!oMǢc\8'Q^UKc՝(kHNl,M+3^,y!#i*ZP.RA[r,5ղSwЯxyJ dS-'Xq[7yRُO+~~s|-1964ǽ濊)XDYn}! Ufᄪ\8"@)fQ:ض;uk NJFzQ!htǭFJZɤUfX[~n=5V}|] Բ'\T[j^~Y"e]}wD󔍅zIwZ}^ȥQȚB;hڹ' Z X iN!\'@wa]8jfF~4 ̣H auҢVDjGxuZHS 4IxέJ C&շKׄ⥁ "(r1є1oYUz(2XЉh(̘=,{KlUMNkdRKYUdcXU;gVŖlzK ؉`W:pKTRD/drm’HgA9EiB43 ;Ed(>_]^UK ')Vtf0((!9ZXX'o˴M["է[iP[6nC4ϊ--IrM~OLኛ>͗zՠtki50X3~,Ndsͮ|PZ?lAyMg9)ʽ"|rF+5#\nƋwmV: (߶ǞS$΁c?:gJgx5gd 4Ȗ N2כ6A߄;$Uߜs 4tJHt|ir~6Fq&D* &*XJSǟ67bSZeB퓷( $#-FvC*iwWW9Iwem$Iz$GdfdаF==()Ӣ({0}#IY*U4a"EVwdFF,8֕!|H 2iJJL+tT̐t&$tՅRFvl$wQ/ܼNm1vd#Nu- 3p%*)\y܏Gm"s晁kD\RDx?Q*Dk3klzWwZQ<1iڬȮq.>'6l2f'=.FU$K,`Mp ku:6) h=:t,ZV=E 9P~_%ZtUr)NkIoɊ‚$OuεS.6᳞@A6Ibä(*`gu8S|+kw2H͝0d@z^\dw&J\[^)ZYO{dȗ])aE=/J\5Θ'E{X61,TJ,z$B-LztXṘW_{V Ѥl"I8lnr<$ °8Am(rV/"-\5();Ƥ$ bwԔ%\ГOr-}b3]2)Ĵ ed:%ayt2'I?_8|$ p] R^ ro hRf~wƙdɻ,\eu,) FY}7$srou3ۍF 6:4N(HbwՀG2G\ߒo{K !G!: l^M"r-7 Ӟ^++PǾ#+B?6}eCkN- m6d)%Wd rÞlNIcG2I陮-'rz"#G1~St`BpnfMo6Bס2{dM.95iŀ!vL($mo)3{ uOŸ67mَe)(@W(Co{tnޣfIph .3~ú#=pCο;%@8|XbKkӋQ#3x矔TV0ʈuku_~]d3i\Zz2qH݈͗R~r hr$w8Gy>P`T=YphBRY +z(89au'-77zp"٬FS +1, t!]")xu<v(hlh/sL~t.ileCF;%b="(+S=b1U^ItK:Ɇ]MF ,:ZCm켉:ʹ:7RCj y$[,ʳ$CnE T)T_0 P_] RR" fU)Ұl}3 E+y[H@S')JL$c}q2o8DYe#ID1/Vg#w(K чսęQ ݼ¡;&yZמD5kZS.i ,_Œo҃~kosrbU?uK_|߭Xf5E%29o^ߐxr|H+pKV% t%DUEt /Zb w6QSVܰvoET)D˃l~:Zs(zS9*9QcHM Uږɫv!R 2{IГ޷4 ; ڂ%*O~G2~2^Mpk^-~0I1;39A\}NW1Z9G}"⧃2;=\@+ڣo9HyzANqrĎ?X|>za[mSmS?qX; 'Oɑn1k +wMhV7:</cem㳓ҥ+&I^u0ѻ\ A"<GW#OPNiye)ԒƩݹwK=-]R(qr8}6cٱ e u֜q{|tZsrupnsomhڬglvpxїRѳ/MĦgǵ7 Z=^yi*㌪ypM/gDj1U{S?(Zͳ4wRopqaZcZABf,Uh Ru ryDR0H,I+^Ct) ^bDr.2&m|9Z8M LΛ-;K'~rpٙمVCO74Aɥ$&eUZD-Čd}*uYDiOh+;VO*ʪβOm}YnmV 5|.#K[cYW#]k9[d}v#<[꾿Ms\16(1Zѷۇ~=n퇃㳣d{}O'w\þ9ׇ?916q*}sԴmv8n~Vghh_yW[f(: &V8tӗ|kxCoqbg|8hh\J|шeKbPCekU|cgֆmkGMDB׏O\^h7KN;O;|p'wZ`k-X_+WxQ͙>qa 4]u^^?$JZV59lt'7܂k1`C)UJ 2 \b>|5G" =?Ҵ~!6"c8Zld=/L%@hg|\ QI _w]JUjj~1${ip6Φo?W;|A*}tYXQqnCv,wi8 W^Ѳ}|1Cz9Y{gi`kuq&c3d=퇒!9:6dn]uh-JA?)x,fĚزyyȨ-11@wɵgS{.3t JEoNOYmhms-&3}x9*f] (bAEddlgpݥ֢V&'nxOɁO9S~<^N(h>yB̜!YL*ϩTk-L_9(I +62R᤭dWwEY; e=;s6j1paR@n>qށpm3;W0")|S[!aDDvqk9 U1JIl-@#B9{]!;+/]\BN]h5aaG,'V[W\e=+IJ֮\;+Q(MɃ4{/9mF3( `Ƭ0bet)(o=XcJnβ;8"4Ӧ"85ơOdZ-s!'lc1< #Vc1b Bv`H@' ,iC t8e[%]q=~d6?\<>iw1Db4>|+TcOY#,AX`b2eOD@[=a?(jQ6!9f!3P;GH;*kgwT+= !r#) %lɎ䐓W$lh>YqfTD(&pq2lB!M5)LݎL Qj/+v\}`rc9k B 6[&dQ{.E)ɢGx+o7~w؏tbHTh+)Uw VjS2uSs}84ci&ePQ ,JP <+)<N]}e-$aAwq^,;7ҒtGLR$u 'kOW!ɤX' rdb]0V\Z""Zp)˹yN`Kİ qǭhBZ@(&!sKշR's6:+ 3:rry$Y'$Z:' 1Ե>=qBa/ 3!q!̳j ) ѓ Z}Ɨ_Y;\GB!Ȏڻ6n%WXz )qwaV]r9ON0,)$:}CR$ 13LbGGFw/@k?]~]+n+29Ynkh5! e1a͎ӑIOg@ {n<[P6쏃8*N;+Ő>\wLT w* ATF /@ó&J_NYg,T4?_\j[̯l䔤k2s "K02[Fj3DT+4mF+׮Mf &Mb`Fs>䒑p |h6K(E1ʏ1̊arihWF3*ZPm'kො] cCdP)Q6۬$Dӥ'"[QFY2&'|?~5`ryh3 Ո#'+0jVߓmH2NFNBJjDc^u#&5#=kR_ U7Y2ˍ9U8)}N)UkvX]{Ka=ܯH%M|HR9SjM*{R6ͮub5'.K[??L / IJ!2[CȜ+qn}oto ժ&iLTwySKG/p]f$ [#GG0[02#1AvEfm,1&rBz7Dy#U :6m.jZsq&*ܶ!]@Vd0ʣ&B 4ȜS6Fz$c) {qd{|=>x 'ʠrMW_n>{$?_y}d<.nv{e =+ʼn bpk'3#AAHF (bp`O;|W׶)Rj$H>N 8g?}|wO2LqHd]pXY3ŠEh!hRǣEznSrFɎGoo40|GHJ M5ʖ;ܖ^6|LK['kE ~ .1lzoY3Uj1=# wMa f!Nj gU$*1L > Y=Ϙd0k4NV L5cf蒨j"}> ?rPqCxD>3g-'$EFWX]PVa7e0Qɂbf~6W,7޿Ln|WN gQ>Y ]?Ӱ+ys7c3-=s? ߙnjp,<ov\<ƫn03_PE0K[IttlŀtD_J+/qWaī|CI3>rdS آ[YhZqbh7Cjf\ ̔Nlh7͚,x,0"*mvboA1 /TZQaz qNORجi=8 # eêV >ʎlMoN~vwۏbr C??x7̓|{G1kcW ޕZ?17Mw{;'%ض$&Nms)n'9>rQGR ymF2F;6Y̼_Od΋~/3&{fµ}:l&(&Y;[Sh!t|˲H& TDH=c h 6ܯvSMcݩhg: JԞZ,?՘3 yӎS vOoL,pfjA70TX8MgA1#S$nF\`:o33K;^ۦf3pR>{hCJ;ǫǸxp8O[z3p:ҒRok 7s)n'!pLjmzϺbKguqpN~=zܰn /օoCL82=ع\z޹\& Li\Z<Ԅ|ߟjEXA& I3rz"{%WR-"CzղR}@TĀYQoCnjE4!'I:U'=cH/ФRy:cԱnM2ҙuKnMhȉhNqsӺJSZTĨNu?;$2֭ 9q]өfevBi+/n97lӑ]X \\&Tk}+ >҈)~/)ivԱ @Q=uܮr#Nn5|ڝc9ĸ"pX^LGQ*1cGkA"Wf. &ܯv3׷+{bl#Xyu̸i'VbkorTJBd#eQ6ץr섴^|HχC= ap$rʹ2YQ)1E+5y!,{Ǩ%P\!̮,d:э .?qBYi^:+M?81W:Ekr,4h'=RW}oGo1c$+$sUP3XA$`1g6D0Iy3 VyH(d@ )d3i<^5\Joמi@RU٣#[/"j1"F7Hɘs'!Ѫմ<~Ʒ\oor9K2A֦S1F ;DR5kǝ1g֌q&*]DIˆˠP50Ģ?Lp,y yF=sqUإJP IQA  RI=Pz (xd,z0T:ߚuu, ?TczH-$NKK5g-=:-!Twybvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004707004515140346171017706 0ustar rootrootFeb 03 09:10:07 crc systemd[1]: Starting Kubernetes Kubelet... Feb 03 09:10:07 crc restorecon[4674]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 03 09:10:07 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 03 09:10:08 crc restorecon[4674]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 03 09:10:09 crc kubenswrapper[4756]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 03 09:10:09 crc kubenswrapper[4756]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 03 09:10:09 crc kubenswrapper[4756]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 03 09:10:09 crc kubenswrapper[4756]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 03 09:10:09 crc kubenswrapper[4756]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 03 09:10:09 crc kubenswrapper[4756]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.336949 4756 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.341929 4756 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.341989 4756 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342004 4756 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342014 4756 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342023 4756 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342032 4756 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342039 4756 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342047 4756 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342055 4756 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342078 4756 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342089 4756 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342099 4756 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342107 4756 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342115 4756 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342123 4756 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342131 4756 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342139 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342147 4756 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342155 4756 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342162 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342171 4756 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342179 4756 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342187 4756 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342197 4756 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342208 4756 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342216 4756 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342225 4756 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342233 4756 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342241 4756 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342249 4756 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342256 4756 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342265 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342273 4756 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342281 4756 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342288 4756 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342296 4756 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342305 4756 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342313 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342321 4756 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342329 4756 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342339 4756 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342349 4756 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342358 4756 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342367 4756 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342375 4756 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342385 4756 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342392 4756 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342402 4756 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342412 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342422 4756 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342430 4756 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342439 4756 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342473 4756 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342482 4756 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342491 4756 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342499 4756 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342507 4756 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342515 4756 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342525 4756 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342536 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342556 4756 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342565 4756 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342573 4756 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342582 4756 feature_gate.go:330] unrecognized feature gate: Example Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342589 4756 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342597 4756 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342604 4756 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342612 4756 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342620 4756 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342628 4756 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.342636 4756 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342771 4756 flags.go:64] FLAG: --address="0.0.0.0" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342787 4756 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342801 4756 flags.go:64] FLAG: --anonymous-auth="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342813 4756 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342824 4756 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342834 4756 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342845 4756 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342856 4756 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342865 4756 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342874 4756 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342884 4756 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342894 4756 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342903 4756 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342912 4756 flags.go:64] FLAG: --cgroup-root="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342921 4756 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342929 4756 flags.go:64] FLAG: --client-ca-file="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342938 4756 flags.go:64] FLAG: --cloud-config="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342947 4756 flags.go:64] FLAG: --cloud-provider="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342955 4756 flags.go:64] FLAG: --cluster-dns="[]" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342965 4756 flags.go:64] FLAG: --cluster-domain="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342974 4756 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342982 4756 flags.go:64] FLAG: --config-dir="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.342991 4756 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343001 4756 flags.go:64] FLAG: --container-log-max-files="5" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343012 4756 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343023 4756 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343033 4756 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343043 4756 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343053 4756 flags.go:64] FLAG: --contention-profiling="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343062 4756 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343071 4756 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343080 4756 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343089 4756 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343102 4756 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343114 4756 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343127 4756 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343138 4756 flags.go:64] FLAG: --enable-load-reader="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343149 4756 flags.go:64] FLAG: --enable-server="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343160 4756 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343175 4756 flags.go:64] FLAG: --event-burst="100" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343186 4756 flags.go:64] FLAG: --event-qps="50" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343198 4756 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343210 4756 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343221 4756 flags.go:64] FLAG: --eviction-hard="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343234 4756 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343243 4756 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343252 4756 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343261 4756 flags.go:64] FLAG: --eviction-soft="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343270 4756 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343279 4756 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343288 4756 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343297 4756 flags.go:64] FLAG: --experimental-mounter-path="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343306 4756 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343315 4756 flags.go:64] FLAG: --fail-swap-on="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343324 4756 flags.go:64] FLAG: --feature-gates="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343335 4756 flags.go:64] FLAG: --file-check-frequency="20s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343344 4756 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343355 4756 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343366 4756 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343379 4756 flags.go:64] FLAG: --healthz-port="10248" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343391 4756 flags.go:64] FLAG: --help="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343404 4756 flags.go:64] FLAG: --hostname-override="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343415 4756 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343427 4756 flags.go:64] FLAG: --http-check-frequency="20s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343439 4756 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343479 4756 flags.go:64] FLAG: --image-credential-provider-config="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343491 4756 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343514 4756 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343526 4756 flags.go:64] FLAG: --image-service-endpoint="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343537 4756 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343549 4756 flags.go:64] FLAG: --kube-api-burst="100" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343560 4756 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343572 4756 flags.go:64] FLAG: --kube-api-qps="50" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343584 4756 flags.go:64] FLAG: --kube-reserved="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343593 4756 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343603 4756 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343616 4756 flags.go:64] FLAG: --kubelet-cgroups="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343628 4756 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343640 4756 flags.go:64] FLAG: --lock-file="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343652 4756 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343664 4756 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343676 4756 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343693 4756 flags.go:64] FLAG: --log-json-split-stream="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343704 4756 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343715 4756 flags.go:64] FLAG: --log-text-split-stream="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343727 4756 flags.go:64] FLAG: --logging-format="text" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343739 4756 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343751 4756 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343762 4756 flags.go:64] FLAG: --manifest-url="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343773 4756 flags.go:64] FLAG: --manifest-url-header="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343788 4756 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343800 4756 flags.go:64] FLAG: --max-open-files="1000000" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343813 4756 flags.go:64] FLAG: --max-pods="110" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343824 4756 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343835 4756 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343847 4756 flags.go:64] FLAG: --memory-manager-policy="None" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343858 4756 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343872 4756 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343884 4756 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343896 4756 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343919 4756 flags.go:64] FLAG: --node-status-max-images="50" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343930 4756 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343942 4756 flags.go:64] FLAG: --oom-score-adj="-999" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343953 4756 flags.go:64] FLAG: --pod-cidr="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343965 4756 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343983 4756 flags.go:64] FLAG: --pod-manifest-path="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.343994 4756 flags.go:64] FLAG: --pod-max-pids="-1" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344006 4756 flags.go:64] FLAG: --pods-per-core="0" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344017 4756 flags.go:64] FLAG: --port="10250" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344028 4756 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344039 4756 flags.go:64] FLAG: --provider-id="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344050 4756 flags.go:64] FLAG: --qos-reserved="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344063 4756 flags.go:64] FLAG: --read-only-port="10255" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344074 4756 flags.go:64] FLAG: --register-node="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344085 4756 flags.go:64] FLAG: --register-schedulable="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344096 4756 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344116 4756 flags.go:64] FLAG: --registry-burst="10" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344128 4756 flags.go:64] FLAG: --registry-qps="5" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344139 4756 flags.go:64] FLAG: --reserved-cpus="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344150 4756 flags.go:64] FLAG: --reserved-memory="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344163 4756 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344175 4756 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344186 4756 flags.go:64] FLAG: --rotate-certificates="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344197 4756 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344207 4756 flags.go:64] FLAG: --runonce="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344218 4756 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344230 4756 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344242 4756 flags.go:64] FLAG: --seccomp-default="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344252 4756 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344263 4756 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344274 4756 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344286 4756 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344297 4756 flags.go:64] FLAG: --storage-driver-password="root" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344311 4756 flags.go:64] FLAG: --storage-driver-secure="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344322 4756 flags.go:64] FLAG: --storage-driver-table="stats" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344333 4756 flags.go:64] FLAG: --storage-driver-user="root" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344345 4756 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344358 4756 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344370 4756 flags.go:64] FLAG: --system-cgroups="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344380 4756 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344398 4756 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344408 4756 flags.go:64] FLAG: --tls-cert-file="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344420 4756 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344433 4756 flags.go:64] FLAG: --tls-min-version="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344478 4756 flags.go:64] FLAG: --tls-private-key-file="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344490 4756 flags.go:64] FLAG: --topology-manager-policy="none" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344501 4756 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344512 4756 flags.go:64] FLAG: --topology-manager-scope="container" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344524 4756 flags.go:64] FLAG: --v="2" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344538 4756 flags.go:64] FLAG: --version="false" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344552 4756 flags.go:64] FLAG: --vmodule="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344565 4756 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.344577 4756 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344814 4756 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344826 4756 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344835 4756 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344843 4756 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344853 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344862 4756 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344870 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344878 4756 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344887 4756 feature_gate.go:330] unrecognized feature gate: Example Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344895 4756 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344903 4756 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344912 4756 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344920 4756 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344928 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344935 4756 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344943 4756 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344952 4756 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344960 4756 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344968 4756 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344976 4756 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344984 4756 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.344992 4756 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345000 4756 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345008 4756 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345015 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345023 4756 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345031 4756 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345039 4756 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345046 4756 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345054 4756 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345062 4756 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345069 4756 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345077 4756 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345085 4756 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345093 4756 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345101 4756 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345108 4756 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345117 4756 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345124 4756 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345132 4756 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345139 4756 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345147 4756 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345155 4756 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345166 4756 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345176 4756 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345186 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345194 4756 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345203 4756 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345211 4756 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345219 4756 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345226 4756 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345236 4756 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345247 4756 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345256 4756 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345265 4756 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345273 4756 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345284 4756 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345293 4756 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345301 4756 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345309 4756 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345317 4756 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345325 4756 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345332 4756 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345341 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345348 4756 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345356 4756 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345363 4756 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345371 4756 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345379 4756 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345387 4756 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.345394 4756 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.345418 4756 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.362432 4756 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.362536 4756 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362681 4756 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362696 4756 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362706 4756 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362715 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362724 4756 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362734 4756 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362743 4756 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362752 4756 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362765 4756 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362775 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362784 4756 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362792 4756 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362801 4756 feature_gate.go:330] unrecognized feature gate: Example Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362811 4756 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362821 4756 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362831 4756 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362840 4756 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362848 4756 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362856 4756 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362864 4756 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362872 4756 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362880 4756 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362888 4756 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362895 4756 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362903 4756 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362913 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362920 4756 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362929 4756 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362936 4756 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362945 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362954 4756 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362962 4756 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362971 4756 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362979 4756 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362987 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.362995 4756 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363007 4756 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363018 4756 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363028 4756 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363036 4756 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363046 4756 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363054 4756 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363062 4756 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363070 4756 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363078 4756 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363086 4756 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363094 4756 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363102 4756 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363110 4756 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363121 4756 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363130 4756 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363139 4756 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363150 4756 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363159 4756 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363168 4756 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363176 4756 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363185 4756 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363195 4756 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363204 4756 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363213 4756 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363222 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363232 4756 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363244 4756 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363253 4756 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363262 4756 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363271 4756 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363279 4756 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363288 4756 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363296 4756 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363303 4756 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363312 4756 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.363326 4756 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363596 4756 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363612 4756 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363620 4756 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363628 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363637 4756 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363650 4756 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363660 4756 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363669 4756 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363678 4756 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363686 4756 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363695 4756 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363702 4756 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363710 4756 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363717 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363725 4756 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363733 4756 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363741 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363748 4756 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363756 4756 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363764 4756 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363772 4756 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363780 4756 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363788 4756 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363795 4756 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363804 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363814 4756 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363824 4756 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363833 4756 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363841 4756 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363849 4756 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363856 4756 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363865 4756 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363872 4756 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363880 4756 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363903 4756 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363911 4756 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363922 4756 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363932 4756 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363941 4756 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363949 4756 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363957 4756 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363966 4756 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363974 4756 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363982 4756 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363990 4756 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.363998 4756 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364008 4756 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364018 4756 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364027 4756 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364036 4756 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364044 4756 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364052 4756 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364061 4756 feature_gate.go:330] unrecognized feature gate: Example Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364070 4756 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364078 4756 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364086 4756 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364094 4756 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364102 4756 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364110 4756 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364120 4756 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364128 4756 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364135 4756 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364143 4756 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364151 4756 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364159 4756 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364169 4756 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364179 4756 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364187 4756 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364196 4756 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364204 4756 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.364211 4756 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.364224 4756 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.365712 4756 server.go:940] "Client rotation is on, will bootstrap in background" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.374514 4756 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.374685 4756 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.377634 4756 server.go:997] "Starting client certificate rotation" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.377682 4756 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.379643 4756 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-13 21:38:16.145356788 +0000 UTC Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.379745 4756 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.406629 4756 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.408916 4756 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.409844 4756 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.426477 4756 log.go:25] "Validated CRI v1 runtime API" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.469547 4756 log.go:25] "Validated CRI v1 image API" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.473012 4756 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.480131 4756 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-03-09-05-55-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.480190 4756 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.515069 4756 manager.go:217] Machine: {Timestamp:2026-02-03 09:10:09.508965091 +0000 UTC m=+0.659432546 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8417792e-76c4-4111-84f0-a989146d0caa BootID:5e8183e3-1710-4801-877d-2bd18fb91137 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:98:71:37 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:98:71:37 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:c1:ed:1b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:bc:9b:91 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:af:82:a4 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:90:27:40 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:92:50:97:66:ca:11 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9e:37:93:b9:df:cd Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.515530 4756 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.515738 4756 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.516515 4756 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.516867 4756 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.516926 4756 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.517353 4756 topology_manager.go:138] "Creating topology manager with none policy" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.517385 4756 container_manager_linux.go:303] "Creating device plugin manager" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.518174 4756 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.518251 4756 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.519635 4756 state_mem.go:36] "Initialized new in-memory state store" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.519803 4756 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.523752 4756 kubelet.go:418] "Attempting to sync node with API server" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.523796 4756 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.523874 4756 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.523899 4756 kubelet.go:324] "Adding apiserver pod source" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.523920 4756 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.529398 4756 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.530978 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.530984 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.531080 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.531115 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.531814 4756 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.533576 4756 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535799 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535849 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535867 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535882 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535906 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535923 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535938 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535962 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.535979 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.536187 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.536242 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.536261 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.538393 4756 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.539155 4756 server.go:1280] "Started kubelet" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.539312 4756 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.539510 4756 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.540791 4756 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 03 09:10:09 crc systemd[1]: Started Kubernetes Kubelet. Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.541775 4756 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.542401 4756 server.go:460] "Adding debug handlers to kubelet server" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.544084 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.544188 4756 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.544423 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 10:05:04.06481323 +0000 UTC Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.544574 4756 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.544664 4756 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.544678 4756 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.544760 4756 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.545545 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.545632 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.546116 4756 factory.go:55] Registering systemd factory Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.546150 4756 factory.go:221] Registration of the systemd container factory successfully Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.550694 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="200ms" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.551530 4756 factory.go:153] Registering CRI-O factory Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.551566 4756 factory.go:221] Registration of the crio container factory successfully Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.551671 4756 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.551721 4756 factory.go:103] Registering Raw factory Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.551746 4756 manager.go:1196] Started watching for new ooms in manager Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.552770 4756 manager.go:319] Starting recovery of all containers Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.557011 4756 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.121:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1890b17e4b035d82 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-03 09:10:09.539112322 +0000 UTC m=+0.689579737,LastTimestamp:2026-02-03 09:10:09.539112322 +0000 UTC m=+0.689579737,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.569280 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.569415 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.569440 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.569485 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.569505 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.569934 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.569956 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570016 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570041 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570060 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570079 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570100 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570120 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570147 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570167 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570189 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570208 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570228 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570249 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570269 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570290 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570317 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570337 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570388 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570411 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570432 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570478 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570501 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570525 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570547 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570566 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570616 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570638 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570658 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570679 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570700 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570719 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570738 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570758 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570819 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570838 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570860 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570879 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570899 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570919 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570940 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570960 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.570984 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571005 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571024 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571045 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571065 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571112 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571136 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571158 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571183 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571205 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571226 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571247 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571294 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571315 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571337 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571357 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571379 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571399 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571420 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571440 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571488 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571509 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571530 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571550 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571571 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571591 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571610 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571631 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571653 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571675 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571694 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571714 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571735 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571754 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571792 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571810 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571830 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571849 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571870 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571892 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571913 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571939 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571961 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.571982 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572002 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572022 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572041 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572062 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572084 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572103 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572129 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572149 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572169 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572191 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572213 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572231 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572275 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572376 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572400 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.572424 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574689 4756 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574735 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574759 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574783 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574807 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574827 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574856 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574881 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574906 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574928 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574947 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574966 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.574984 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575004 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575022 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575041 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575059 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575078 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575096 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575116 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575136 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575154 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575175 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575194 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575213 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575231 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575249 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575269 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575288 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575306 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575323 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575343 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575361 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575381 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575401 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575419 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575439 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575485 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575519 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575540 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575559 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575578 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575598 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575619 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575640 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575657 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575676 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575694 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575713 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575731 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575750 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575769 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575789 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575810 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575829 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575850 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575869 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575891 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575912 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575931 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575949 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575967 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.575985 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576002 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576019 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576040 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576059 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576078 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576098 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576116 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576134 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576152 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576172 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576193 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576211 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576230 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576247 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576267 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576285 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576303 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576321 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576341 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576361 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576378 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576399 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576418 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576436 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576475 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576496 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576516 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576535 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576555 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576574 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576595 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576614 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576636 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576659 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576678 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576696 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576714 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576736 4756 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576754 4756 reconstruct.go:97] "Volume reconstruction finished" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.576768 4756 reconciler.go:26] "Reconciler: start to sync state" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.578264 4756 manager.go:324] Recovery completed Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.588193 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.592016 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.592072 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.592089 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.593359 4756 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.593387 4756 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.593426 4756 state_mem.go:36] "Initialized new in-memory state store" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.611291 4756 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.612820 4756 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.612875 4756 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.612913 4756 kubelet.go:2335] "Starting kubelet main sync loop" Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.612981 4756 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 03 09:10:09 crc kubenswrapper[4756]: W0203 09:10:09.614258 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.614382 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.620913 4756 policy_none.go:49] "None policy: Start" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.623396 4756 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.623437 4756 state_mem.go:35] "Initializing new in-memory state store" Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.645349 4756 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.685910 4756 manager.go:334] "Starting Device Plugin manager" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.686007 4756 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.686834 4756 server.go:79] "Starting device plugin registration server" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.687574 4756 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.687831 4756 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.688011 4756 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.688138 4756 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.688149 4756 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.695814 4756 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.714618 4756 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.714817 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.716509 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.716552 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.716567 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.716797 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.716991 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.717039 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.717731 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.717761 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.717778 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.717788 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.717810 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.717821 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.717953 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.718164 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.718233 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.718891 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.718919 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.718948 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.719067 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.719202 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.719238 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720638 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720739 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720751 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720642 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720807 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720818 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720711 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720902 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720921 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.720975 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721034 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721083 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721691 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721716 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721732 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721885 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721942 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721909 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.721959 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.722663 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.722684 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.722695 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.752033 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="400ms" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.779754 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.779798 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.779818 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.779835 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.779904 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.779966 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780001 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780019 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780035 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780071 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780103 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780123 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780145 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780164 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.780185 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.788703 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.789659 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.789692 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.789702 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.789728 4756 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.790034 4756 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.121:6443: connect: connection refused" node="crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881433 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881531 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881552 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881573 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881591 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881615 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881632 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881648 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881663 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881719 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881752 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881746 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881789 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881752 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881664 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881788 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881672 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881746 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881914 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881946 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881972 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.881996 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.882015 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.882020 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.882035 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.882053 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.882062 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.882091 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.882096 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.882184 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.990478 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.992102 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.992177 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.992188 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:09 crc kubenswrapper[4756]: I0203 09:10:09.992229 4756 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 03 09:10:09 crc kubenswrapper[4756]: E0203 09:10:09.992677 4756 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.121:6443: connect: connection refused" node="crc" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.054954 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.063185 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.079719 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.098006 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.102985 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:10 crc kubenswrapper[4756]: W0203 09:10:10.110179 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-e6020add8787e095aa8a62f4b666f72e92c4bc5672b89ed27c09bbdb730ca032 WatchSource:0}: Error finding container e6020add8787e095aa8a62f4b666f72e92c4bc5672b89ed27c09bbdb730ca032: Status 404 returned error can't find the container with id e6020add8787e095aa8a62f4b666f72e92c4bc5672b89ed27c09bbdb730ca032 Feb 03 09:10:10 crc kubenswrapper[4756]: W0203 09:10:10.110527 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-77a5ca01aeb4ad0002ffad6cfaacfa6259100923c904e481135a460e8a8eb506 WatchSource:0}: Error finding container 77a5ca01aeb4ad0002ffad6cfaacfa6259100923c904e481135a460e8a8eb506: Status 404 returned error can't find the container with id 77a5ca01aeb4ad0002ffad6cfaacfa6259100923c904e481135a460e8a8eb506 Feb 03 09:10:10 crc kubenswrapper[4756]: W0203 09:10:10.115834 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-0cc970b93ac2b6500d296b5550eb111797f5f5c53094311f13182ff5a91bf349 WatchSource:0}: Error finding container 0cc970b93ac2b6500d296b5550eb111797f5f5c53094311f13182ff5a91bf349: Status 404 returned error can't find the container with id 0cc970b93ac2b6500d296b5550eb111797f5f5c53094311f13182ff5a91bf349 Feb 03 09:10:10 crc kubenswrapper[4756]: W0203 09:10:10.123744 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f713598e4b89ae35111aeedca23d2e6ea025b593236a1c9aa543967ac36436ce WatchSource:0}: Error finding container f713598e4b89ae35111aeedca23d2e6ea025b593236a1c9aa543967ac36436ce: Status 404 returned error can't find the container with id f713598e4b89ae35111aeedca23d2e6ea025b593236a1c9aa543967ac36436ce Feb 03 09:10:10 crc kubenswrapper[4756]: W0203 09:10:10.131530 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7771af13e99755568fc1abf0a3306499d005a11b2a1bb7ec1798d89037a70103 WatchSource:0}: Error finding container 7771af13e99755568fc1abf0a3306499d005a11b2a1bb7ec1798d89037a70103: Status 404 returned error can't find the container with id 7771af13e99755568fc1abf0a3306499d005a11b2a1bb7ec1798d89037a70103 Feb 03 09:10:10 crc kubenswrapper[4756]: E0203 09:10:10.152895 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="800ms" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.393437 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.394778 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.395099 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.395114 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.395139 4756 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 03 09:10:10 crc kubenswrapper[4756]: E0203 09:10:10.395605 4756 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.121:6443: connect: connection refused" node="crc" Feb 03 09:10:10 crc kubenswrapper[4756]: W0203 09:10:10.489075 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:10 crc kubenswrapper[4756]: E0203 09:10:10.489179 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.543235 4756 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.545320 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 01:32:04.753711379 +0000 UTC Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.618646 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e6020add8787e095aa8a62f4b666f72e92c4bc5672b89ed27c09bbdb730ca032"} Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.620955 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7771af13e99755568fc1abf0a3306499d005a11b2a1bb7ec1798d89037a70103"} Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.622315 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f713598e4b89ae35111aeedca23d2e6ea025b593236a1c9aa543967ac36436ce"} Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.623677 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0cc970b93ac2b6500d296b5550eb111797f5f5c53094311f13182ff5a91bf349"} Feb 03 09:10:10 crc kubenswrapper[4756]: I0203 09:10:10.625132 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"77a5ca01aeb4ad0002ffad6cfaacfa6259100923c904e481135a460e8a8eb506"} Feb 03 09:10:10 crc kubenswrapper[4756]: W0203 09:10:10.654016 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:10 crc kubenswrapper[4756]: E0203 09:10:10.654092 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:10 crc kubenswrapper[4756]: W0203 09:10:10.715527 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:10 crc kubenswrapper[4756]: E0203 09:10:10.715611 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:10 crc kubenswrapper[4756]: E0203 09:10:10.954576 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="1.6s" Feb 03 09:10:11 crc kubenswrapper[4756]: W0203 09:10:11.061548 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:11 crc kubenswrapper[4756]: E0203 09:10:11.061629 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.195653 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.197345 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.197373 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.197384 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.197422 4756 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 03 09:10:11 crc kubenswrapper[4756]: E0203 09:10:11.197901 4756 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.121:6443: connect: connection refused" node="crc" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.499709 4756 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 03 09:10:11 crc kubenswrapper[4756]: E0203 09:10:11.500859 4756 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.543232 4756 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.545400 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 14:52:25.000041629 +0000 UTC Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.629756 4756 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="01bba55be891fea202cda90272e83cf96b442597adfac7bc37d71bc9685ddc78" exitCode=0 Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.629816 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"01bba55be891fea202cda90272e83cf96b442597adfac7bc37d71bc9685ddc78"} Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.629921 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.631246 4756 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4" exitCode=0 Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.631286 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4"} Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.631346 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.631359 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.631366 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.631581 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.636727 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.636759 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.636768 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.636768 4756 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57" exitCode=0 Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.636837 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57"} Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.636851 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.638058 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.638224 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.638356 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.640965 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8"} Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.640993 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451"} Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.641004 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1"} Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.641014 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81"} Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.641391 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.642273 4756 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="80d73bf3db31948c8a05591d65a087d4e068bd35d6728fc117e1dd32f6570dbe" exitCode=0 Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.642304 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"80d73bf3db31948c8a05591d65a087d4e068bd35d6728fc117e1dd32f6570dbe"} Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.642390 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.642710 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.642746 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.642757 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.643467 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.643484 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.643492 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.644777 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.646308 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.646347 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:11 crc kubenswrapper[4756]: I0203 09:10:11.646365 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:12 crc kubenswrapper[4756]: W0203 09:10:12.401792 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:12 crc kubenswrapper[4756]: E0203 09:10:12.401866 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:12 crc kubenswrapper[4756]: W0203 09:10:12.507827 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:12 crc kubenswrapper[4756]: E0203 09:10:12.507909 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.542705 4756 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.545851 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 13:39:32.463016837 +0000 UTC Feb 03 09:10:12 crc kubenswrapper[4756]: E0203 09:10:12.555339 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="3.2s" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.648503 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ab6bc4d7606d924bcf2bfa638abe04d8ec4c4186e9cec487bdd13a75317c8afd"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.648569 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.648589 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.648601 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.648613 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.648610 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.651949 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.651995 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.652009 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.654666 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.655082 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.655118 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.655131 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.655690 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.655741 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.655774 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.657061 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2cb345ac4f3c06479f5dffe90991140eb445c45550b2958c95e38e23451f9c3e"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.657088 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.658207 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.658258 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.658271 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.659132 4756 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bc5f582bd64c86e38d423f96949ae1daaf5ea41ded446aeee26a4f7d5e779c39" exitCode=0 Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.659241 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.659384 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.659415 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bc5f582bd64c86e38d423f96949ae1daaf5ea41ded446aeee26a4f7d5e779c39"} Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.659860 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.659887 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.659899 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.660191 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.660217 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.660229 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.798247 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.801371 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.801486 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.801501 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:12 crc kubenswrapper[4756]: I0203 09:10:12.801531 4756 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 03 09:10:12 crc kubenswrapper[4756]: E0203 09:10:12.802125 4756 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.121:6443: connect: connection refused" node="crc" Feb 03 09:10:12 crc kubenswrapper[4756]: W0203 09:10:12.886624 4756 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.121:6443: connect: connection refused Feb 03 09:10:12 crc kubenswrapper[4756]: E0203 09:10:12.886724 4756 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.121:6443: connect: connection refused" logger="UnhandledError" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.381828 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.390318 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.546330 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 22:53:09.84761524 +0000 UTC Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664575 4756 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bb39287219460ee48d5440e61b584ed984f1b2dc31335ae0ac73f85725220ee5" exitCode=0 Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664688 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bb39287219460ee48d5440e61b584ed984f1b2dc31335ae0ac73f85725220ee5"} Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664726 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664747 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664766 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664824 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664880 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664834 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664959 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.664896 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.665924 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.665955 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.665963 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666142 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666169 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666178 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666626 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666654 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666664 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666698 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666706 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666725 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666738 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666747 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:13 crc kubenswrapper[4756]: I0203 09:10:13.666762 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.342438 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.453186 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.546756 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 06:00:47.202832614 +0000 UTC Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674185 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674232 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674164 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c96dc6e584323741b447735a2e10d570b8418d3e553630745a7b9d6826d457f3"} Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674323 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4b9bf398273a6803d106a03e473958c126182e4d70d02d77d1d8c62aa33b9c18"} Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674338 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674349 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8dd0f83198ce3d109862b089f7215db6e0c7137a1ff967727eee8f87a245d025"} Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674476 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0e443b4f2cd6ea4e45e78cf81711e53965bc242bf53954c5d3b5ec7552743448"} Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674490 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"edd400658b98d7520ac371563404ddf15eab9d0d5a66fe7dae90bad2147b1661"} Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.674201 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.675430 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.675492 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.675514 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.675936 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.676102 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.676246 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.676389 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.676416 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:14 crc kubenswrapper[4756]: I0203 09:10:14.676427 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.547821 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 02:18:24.672610122 +0000 UTC Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.676310 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.676378 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.677979 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.678014 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.678026 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.677987 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.678082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.678095 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:15 crc kubenswrapper[4756]: I0203 09:10:15.750519 4756 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.003141 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.004821 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.004866 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.004876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.004907 4756 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.548625 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 14:12:17.305761618 +0000 UTC Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.558929 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.559127 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.559175 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.560163 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.560193 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:16 crc kubenswrapper[4756]: I0203 09:10:16.560203 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.411553 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.411756 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.412903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.412953 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.412965 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.548772 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 19:25:22.009930067 +0000 UTC Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.615294 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.680390 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.681506 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.681556 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:17 crc kubenswrapper[4756]: I0203 09:10:17.681569 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:18 crc kubenswrapper[4756]: I0203 09:10:18.549586 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 08:50:22.938403269 +0000 UTC Feb 03 09:10:18 crc kubenswrapper[4756]: I0203 09:10:18.693313 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:18 crc kubenswrapper[4756]: I0203 09:10:18.693692 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:18 crc kubenswrapper[4756]: I0203 09:10:18.695048 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:18 crc kubenswrapper[4756]: I0203 09:10:18.695119 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:18 crc kubenswrapper[4756]: I0203 09:10:18.695157 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:19 crc kubenswrapper[4756]: I0203 09:10:19.271992 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:19 crc kubenswrapper[4756]: I0203 09:10:19.272228 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:19 crc kubenswrapper[4756]: I0203 09:10:19.274021 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:19 crc kubenswrapper[4756]: I0203 09:10:19.274069 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:19 crc kubenswrapper[4756]: I0203 09:10:19.274081 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:19 crc kubenswrapper[4756]: I0203 09:10:19.550522 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 15:02:37.213666004 +0000 UTC Feb 03 09:10:19 crc kubenswrapper[4756]: E0203 09:10:19.696401 4756 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.357325 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.357562 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.361146 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.361252 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.361275 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.364155 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.551498 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 13:35:45.347544397 +0000 UTC Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.687353 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.688740 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.688815 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:20 crc kubenswrapper[4756]: I0203 09:10:20.688833 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:21 crc kubenswrapper[4756]: I0203 09:10:21.552057 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 12:34:21.142974145 +0000 UTC Feb 03 09:10:22 crc kubenswrapper[4756]: I0203 09:10:22.552734 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 20:22:40.132931644 +0000 UTC Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.340977 4756 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.341041 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.348428 4756 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.348546 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.358103 4756 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.358180 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.553034 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 02:49:42.342335029 +0000 UTC Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.698170 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.700369 4756 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ab6bc4d7606d924bcf2bfa638abe04d8ec4c4186e9cec487bdd13a75317c8afd" exitCode=255 Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.700417 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ab6bc4d7606d924bcf2bfa638abe04d8ec4c4186e9cec487bdd13a75317c8afd"} Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.700814 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.701696 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.701758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.701772 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:23 crc kubenswrapper[4756]: I0203 09:10:23.702316 4756 scope.go:117] "RemoveContainer" containerID="ab6bc4d7606d924bcf2bfa638abe04d8ec4c4186e9cec487bdd13a75317c8afd" Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.354720 4756 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]log ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]etcd ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/openshift.io-api-request-count-filter ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/openshift.io-startkubeinformers ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/generic-apiserver-start-informers ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/priority-and-fairness-config-consumer ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/priority-and-fairness-filter ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-apiextensions-informers ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-apiextensions-controllers ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/crd-informer-synced ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-system-namespaces-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-cluster-authentication-info-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-legacy-token-tracking-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-service-ip-repair-controllers ok Feb 03 09:10:24 crc kubenswrapper[4756]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Feb 03 09:10:24 crc kubenswrapper[4756]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/priority-and-fairness-config-producer ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/bootstrap-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/start-kube-aggregator-informers ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/apiservice-status-local-available-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/apiservice-status-remote-available-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/apiservice-registration-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/apiservice-wait-for-first-sync ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/apiservice-discovery-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/kube-apiserver-autoregistration ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]autoregister-completion ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/apiservice-openapi-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: [+]poststarthook/apiservice-openapiv3-controller ok Feb 03 09:10:24 crc kubenswrapper[4756]: livez check failed Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.354848 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.554092 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 10:49:43.073728628 +0000 UTC Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.705370 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.707359 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945"} Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.707498 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.708409 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.708489 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:24 crc kubenswrapper[4756]: I0203 09:10:24.708499 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:25 crc kubenswrapper[4756]: I0203 09:10:25.554719 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 06:49:01.823039709 +0000 UTC Feb 03 09:10:26 crc kubenswrapper[4756]: I0203 09:10:26.554884 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 02:41:34.781322135 +0000 UTC Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.555239 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 11:32:03.221622121 +0000 UTC Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.635968 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.636143 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.637300 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.637344 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.637357 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.647893 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.713222 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.714018 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.714055 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:27 crc kubenswrapper[4756]: I0203 09:10:27.714068 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.349354 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.351691 4756 trace.go:236] Trace[1612427983]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Feb-2026 09:10:18.320) (total time: 10031ms): Feb 03 09:10:28 crc kubenswrapper[4756]: Trace[1612427983]: ---"Objects listed" error: 10031ms (09:10:28.351) Feb 03 09:10:28 crc kubenswrapper[4756]: Trace[1612427983]: [10.031147015s] [10.031147015s] END Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.351720 4756 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.352606 4756 trace.go:236] Trace[1818829062]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Feb-2026 09:10:18.101) (total time: 10250ms): Feb 03 09:10:28 crc kubenswrapper[4756]: Trace[1818829062]: ---"Objects listed" error: 10250ms (09:10:28.352) Feb 03 09:10:28 crc kubenswrapper[4756]: Trace[1818829062]: [10.250906128s] [10.250906128s] END Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.352646 4756 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.354386 4756 trace.go:236] Trace[1217511200]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Feb-2026 09:10:13.661) (total time: 14692ms): Feb 03 09:10:28 crc kubenswrapper[4756]: Trace[1217511200]: ---"Objects listed" error: 14692ms (09:10:28.354) Feb 03 09:10:28 crc kubenswrapper[4756]: Trace[1217511200]: [14.692504606s] [14.692504606s] END Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.354417 4756 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.354395 4756 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.354553 4756 trace.go:236] Trace[210974173]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Feb-2026 09:10:15.783) (total time: 12570ms): Feb 03 09:10:28 crc kubenswrapper[4756]: Trace[210974173]: ---"Objects listed" error: 12570ms (09:10:28.354) Feb 03 09:10:28 crc kubenswrapper[4756]: Trace[210974173]: [12.57070395s] [12.57070395s] END Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.354567 4756 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.355088 4756 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.360529 4756 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.378129 4756 csr.go:261] certificate signing request csr-kzpr9 is approved, waiting to be issued Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.385303 4756 csr.go:257] certificate signing request csr-kzpr9 is issued Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.537038 4756 apiserver.go:52] "Watching apiserver" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.539296 4756 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.540007 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.541003 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.541119 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.541202 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.541391 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.541466 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.541492 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.541742 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.541767 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.541849 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.544131 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.544180 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.544366 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.544502 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.544553 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.544698 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.545685 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.545770 4756 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.546133 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.548962 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555030 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555061 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555079 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555172 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555211 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555230 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555249 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555272 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555296 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555314 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555330 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555346 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555361 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555380 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.555411 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:10:29.055383666 +0000 UTC m=+20.205851041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555459 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555480 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555518 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555543 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555575 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555579 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555604 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555633 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555648 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555658 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555683 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555712 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555738 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555746 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555767 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555791 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555816 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555843 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555868 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555895 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555920 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555946 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555969 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555992 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556014 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556037 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556058 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556086 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556108 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556131 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556153 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556172 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556194 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556218 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556240 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556261 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556283 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556306 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556328 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556352 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556380 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556425 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556465 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556489 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556542 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556569 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556644 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556676 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556704 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556727 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556752 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556778 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556806 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556832 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556854 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556877 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556901 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556928 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556982 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557005 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557038 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557066 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557089 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557111 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557132 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557152 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557173 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557194 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557216 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557237 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557259 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557280 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557304 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557325 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557346 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557368 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557390 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557409 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557430 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557470 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557492 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557515 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557536 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557559 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557581 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557603 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557624 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557646 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557674 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557701 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557723 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557746 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557770 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557791 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557814 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557837 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557860 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557884 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557905 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557930 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557953 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557976 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558000 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558023 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558049 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558074 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558101 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558124 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558149 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558174 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558197 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558221 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558244 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558270 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558295 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558318 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558341 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558367 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558389 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558414 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558456 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558486 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558515 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558542 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558567 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558604 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558629 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558651 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558674 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558702 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558730 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558756 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558779 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558803 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558828 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558851 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558876 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558903 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558926 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558949 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558974 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.558998 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559024 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559048 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559075 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559101 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559125 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559147 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559167 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559187 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559208 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559232 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559256 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559283 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559307 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559330 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559352 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559376 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559401 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559425 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559473 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559501 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559523 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559546 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559571 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559596 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559624 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559651 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559677 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559706 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559731 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559758 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559786 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559810 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559833 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559859 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559885 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559910 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559938 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559956 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559974 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559991 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560008 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560025 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560041 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560085 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560114 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560138 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560157 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560176 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560202 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560221 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560239 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560257 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560275 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560293 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560312 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560330 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560350 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560422 4756 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560468 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560486 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560501 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555793 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.555830 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556116 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556133 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556155 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 17:34:57.001437437 +0000 UTC Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563390 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556281 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563420 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556331 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556561 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556582 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556674 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556707 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.556755 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557158 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557253 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557346 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557350 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557522 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557601 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557830 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.557902 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559660 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559743 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.559944 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560038 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560154 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560359 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560703 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.560893 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.561304 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.561332 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.561648 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.561673 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.561900 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.562038 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.562063 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.562187 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.562364 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.562414 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.562647 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.562704 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563055 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.564060 4756 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.564107 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.564264 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.564341 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.564518 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.564958 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.564966 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563169 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563229 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563341 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563592 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563611 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.565174 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.565194 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.565217 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.565380 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.565420 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.565263 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563120 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.566320 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.566656 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.566936 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.568068 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.570302 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.571234 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.571226 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.571695 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.571761 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.572213 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.572814 4756 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.572874 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:29.072855889 +0000 UTC m=+20.223323334 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.572935 4756 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.572965 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:29.072957072 +0000 UTC m=+20.223424447 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.573099 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.573976 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.574273 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.574785 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.575049 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.575575 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.575706 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.575794 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.575885 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.575927 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.576189 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.576228 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.576384 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.576482 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.576501 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.576918 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.577002 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.577280 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.577349 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.577575 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.577817 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.577905 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.578192 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.578289 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.578391 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.578520 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.579075 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.578714 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.578900 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.579005 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.579420 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.579432 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.579498 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.579655 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.579692 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.579785 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.589086 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.589435 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.589620 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.589834 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.590094 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.590254 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.590473 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.590496 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.590500 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.590521 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.590537 4756 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.590597 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:29.090577699 +0000 UTC m=+20.241045144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.591083 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.592458 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.592790 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.592809 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.592822 4756 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:28 crc kubenswrapper[4756]: E0203 09:10:28.592875 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:29.092856491 +0000 UTC m=+20.243323936 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.593291 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.593314 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.593462 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.594071 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.594137 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.594159 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.594365 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.595158 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.595916 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.596041 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.563655 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.596793 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.597098 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.597319 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.597540 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.607986 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.608016 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.608141 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.608285 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.608307 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.608322 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.608383 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.608620 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.609528 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.609967 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610045 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610071 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610107 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610307 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610331 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610603 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610640 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610655 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.610936 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.611276 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.611360 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.611468 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.612036 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.614033 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.614413 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.614659 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.615124 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.619607 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-cnrvx"] Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.620624 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.622041 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.622762 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-c9rn9"] Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.623107 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.623590 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cnrvx" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.624270 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.624371 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.624692 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.626659 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.626915 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.626974 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.627222 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.627270 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.627816 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.628153 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.628165 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.628173 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.628677 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.629061 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.629079 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.629137 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.629174 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.629669 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.629783 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.629650 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.630326 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.630411 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.631154 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.631016 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.630623 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.632610 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.633991 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.634189 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.634513 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.634564 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.634685 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.636286 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.636363 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.636661 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.637431 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.637705 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.637703 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.637731 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.637752 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.637775 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.639585 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.643369 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.643846 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.644212 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.644631 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.645267 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.653433 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.661134 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.661829 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/748779a5-a5e9-4451-839c-805686b764c5-proxy-tls\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.661892 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/748779a5-a5e9-4451-839c-805686b764c5-rootfs\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.661912 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/748779a5-a5e9-4451-839c-805686b764c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.661936 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.661983 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz2xg\" (UniqueName: \"kubernetes.io/projected/2291b3e7-d0e4-4800-b6bb-9fe15140ab0e-kube-api-access-vz2xg\") pod \"node-resolver-cnrvx\" (UID: \"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\") " pod="openshift-dns/node-resolver-cnrvx" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662007 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wflqj\" (UniqueName: \"kubernetes.io/projected/748779a5-a5e9-4451-839c-805686b764c5-kube-api-access-wflqj\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662092 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662116 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2291b3e7-d0e4-4800-b6bb-9fe15140ab0e-hosts-file\") pod \"node-resolver-cnrvx\" (UID: \"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\") " pod="openshift-dns/node-resolver-cnrvx" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662225 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662273 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662346 4756 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662378 4756 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662399 4756 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662415 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662428 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662439 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662464 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662477 4756 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662489 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662501 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662513 4756 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662524 4756 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662537 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662551 4756 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662565 4756 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662593 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662613 4756 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662626 4756 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662636 4756 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662665 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662679 4756 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662691 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662702 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662714 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662724 4756 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662732 4756 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662740 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662748 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.662758 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663016 4756 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663030 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663042 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663052 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663064 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663076 4756 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663087 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663099 4756 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663111 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663123 4756 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663134 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663146 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663159 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663171 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663182 4756 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663194 4756 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663206 4756 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663220 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663233 4756 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663244 4756 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663256 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663267 4756 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663300 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663313 4756 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663323 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663334 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663345 4756 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663356 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663374 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663386 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663399 4756 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663411 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663423 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663435 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663462 4756 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663506 4756 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663518 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663531 4756 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663542 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663555 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663567 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663579 4756 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663590 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663602 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663615 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663627 4756 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663641 4756 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663652 4756 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663663 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663674 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663686 4756 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663697 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663708 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663725 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663737 4756 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663748 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663759 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663770 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663783 4756 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663795 4756 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663806 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663818 4756 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663829 4756 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663840 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663852 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663890 4756 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663905 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663918 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663929 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663943 4756 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663977 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.663992 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664004 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664014 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664024 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664035 4756 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664047 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664058 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664070 4756 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664084 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664095 4756 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664108 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664119 4756 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664130 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664141 4756 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664152 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664164 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664176 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664187 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664199 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664210 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664221 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664232 4756 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664242 4756 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664254 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664265 4756 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664276 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664287 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664298 4756 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664309 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664320 4756 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664330 4756 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664341 4756 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664351 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664367 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664377 4756 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664389 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664408 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664420 4756 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664431 4756 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664463 4756 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664475 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664486 4756 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664496 4756 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664508 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664519 4756 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664530 4756 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664580 4756 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664593 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664604 4756 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664615 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664627 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664639 4756 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664650 4756 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664661 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664672 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664684 4756 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664696 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664707 4756 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664718 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664730 4756 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664741 4756 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664753 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664764 4756 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664776 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664788 4756 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664801 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664814 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664826 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664838 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664850 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664862 4756 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664874 4756 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664886 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664898 4756 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664910 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664922 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664934 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664947 4756 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664959 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664971 4756 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664982 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.664994 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.665005 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.665018 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.665030 4756 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.665042 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.665053 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.665141 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.665296 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.666404 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.668827 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.675165 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.684069 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.696086 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.706530 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.716016 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.724551 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.734244 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.741311 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.749970 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.758418 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.765974 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/748779a5-a5e9-4451-839c-805686b764c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766023 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz2xg\" (UniqueName: \"kubernetes.io/projected/2291b3e7-d0e4-4800-b6bb-9fe15140ab0e-kube-api-access-vz2xg\") pod \"node-resolver-cnrvx\" (UID: \"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\") " pod="openshift-dns/node-resolver-cnrvx" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766042 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wflqj\" (UniqueName: \"kubernetes.io/projected/748779a5-a5e9-4451-839c-805686b764c5-kube-api-access-wflqj\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766066 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2291b3e7-d0e4-4800-b6bb-9fe15140ab0e-hosts-file\") pod \"node-resolver-cnrvx\" (UID: \"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\") " pod="openshift-dns/node-resolver-cnrvx" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766081 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/748779a5-a5e9-4451-839c-805686b764c5-proxy-tls\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766104 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/748779a5-a5e9-4451-839c-805686b764c5-rootfs\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766124 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766133 4756 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766143 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766185 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/748779a5-a5e9-4451-839c-805686b764c5-rootfs\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766313 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2291b3e7-d0e4-4800-b6bb-9fe15140ab0e-hosts-file\") pod \"node-resolver-cnrvx\" (UID: \"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\") " pod="openshift-dns/node-resolver-cnrvx" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.766739 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/748779a5-a5e9-4451-839c-805686b764c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.769315 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/748779a5-a5e9-4451-839c-805686b764c5-proxy-tls\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.782036 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz2xg\" (UniqueName: \"kubernetes.io/projected/2291b3e7-d0e4-4800-b6bb-9fe15140ab0e-kube-api-access-vz2xg\") pod \"node-resolver-cnrvx\" (UID: \"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\") " pod="openshift-dns/node-resolver-cnrvx" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.782787 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wflqj\" (UniqueName: \"kubernetes.io/projected/748779a5-a5e9-4451-839c-805686b764c5-kube-api-access-wflqj\") pod \"machine-config-daemon-c9rn9\" (UID: \"748779a5-a5e9-4451-839c-805686b764c5\") " pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.854803 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.871166 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.884108 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 03 09:10:28 crc kubenswrapper[4756]: W0203 09:10:28.884535 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-d84769a74a7b23c7c9bf08825995e51ea8b11699dcd016c265979fec0d20c85d WatchSource:0}: Error finding container d84769a74a7b23c7c9bf08825995e51ea8b11699dcd016c265979fec0d20c85d: Status 404 returned error can't find the container with id d84769a74a7b23c7c9bf08825995e51ea8b11699dcd016c265979fec0d20c85d Feb 03 09:10:28 crc kubenswrapper[4756]: W0203 09:10:28.895740 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-da893d8dbe38df8f4b4fa603c805e05d29a06a049b7133446641afd07277c675 WatchSource:0}: Error finding container da893d8dbe38df8f4b4fa603c805e05d29a06a049b7133446641afd07277c675: Status 404 returned error can't find the container with id da893d8dbe38df8f4b4fa603c805e05d29a06a049b7133446641afd07277c675 Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.940937 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.951136 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cnrvx" Feb 03 09:10:28 crc kubenswrapper[4756]: W0203 09:10:28.967853 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod748779a5_a5e9_4451_839c_805686b764c5.slice/crio-7cdd16ad0722165801a8a2cc1a97022b585cef1b1f44d8bc4e0ea9b057e6611a WatchSource:0}: Error finding container 7cdd16ad0722165801a8a2cc1a97022b585cef1b1f44d8bc4e0ea9b057e6611a: Status 404 returned error can't find the container with id 7cdd16ad0722165801a8a2cc1a97022b585cef1b1f44d8bc4e0ea9b057e6611a Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.972965 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-qx7sz"] Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.977626 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-6srkq"] Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.977852 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.978030 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6srkq" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.980975 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.981211 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.981320 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.981409 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.981538 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.981716 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.981801 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 03 09:10:28 crc kubenswrapper[4756]: I0203 09:10:28.991809 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.011379 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.021430 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.031258 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.042521 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.052032 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.058567 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068272 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068386 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-os-release\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068422 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-netns\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068463 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.068522 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:10:30.068414547 +0000 UTC m=+21.218881922 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068562 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cni-binary-copy\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068614 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cnibin\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068632 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-multus-certs\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068651 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-cni-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068669 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-cnibin\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068685 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-kubelet\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068704 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-system-cni-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068720 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-cni-bin\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068738 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ld46\" (UniqueName: \"kubernetes.io/projected/f6a6a991-ccf4-409f-bda5-cabc5788ea88-kube-api-access-6ld46\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068914 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckdq2\" (UniqueName: \"kubernetes.io/projected/89e5d60f-f690-486f-b078-5ae9f98e1f3a-kube-api-access-ckdq2\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068939 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068953 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-conf-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068973 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-os-release\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.068987 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-hostroot\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.069003 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-system-cni-dir\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.069018 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-cni-multus\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.069042 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89e5d60f-f690-486f-b078-5ae9f98e1f3a-cni-binary-copy\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.069056 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-socket-dir-parent\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.069072 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.069096 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-k8s-cni-cncf-io\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.069109 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-daemon-config\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.069122 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-etc-kubernetes\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.078225 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.089518 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.100860 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.110508 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.120511 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.130591 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.141661 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.149218 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.156994 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.164861 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170086 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cnibin\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170131 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-cni-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170162 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-multus-certs\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170185 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-system-cni-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170204 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-cnibin\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170209 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cnibin\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170239 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-kubelet\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170265 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-cni-bin\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170272 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-multus-certs\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170286 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170308 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckdq2\" (UniqueName: \"kubernetes.io/projected/89e5d60f-f690-486f-b078-5ae9f98e1f3a-kube-api-access-ckdq2\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170332 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170355 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ld46\" (UniqueName: \"kubernetes.io/projected/f6a6a991-ccf4-409f-bda5-cabc5788ea88-kube-api-access-6ld46\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170375 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-conf-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170500 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-cni-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170533 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-cni-bin\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170567 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-system-cni-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170591 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-kubelet\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.170649 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-cnibin\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.170675 4756 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.170724 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:30.170710423 +0000 UTC m=+21.321177798 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.171084 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-conf-dir\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.173733 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.173920 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-os-release\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.173961 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-hostroot\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.173990 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-cni-multus\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.173998 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174047 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-system-cni-dir\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174128 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-os-release\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174014 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-system-cni-dir\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174170 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-hostroot\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174198 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-var-lib-cni-multus\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174250 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174312 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89e5d60f-f690-486f-b078-5ae9f98e1f3a-cni-binary-copy\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.174324 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.174347 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.174369 4756 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174385 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-socket-dir-parent\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.174420 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:30.17439921 +0000 UTC m=+21.324866585 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174462 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174500 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174523 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-etc-kubernetes\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174544 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-k8s-cni-cncf-io\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174545 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-socket-dir-parent\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174567 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-daemon-config\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174650 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-os-release\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174708 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cni-binary-copy\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174753 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174827 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-netns\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.174981 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-netns\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.175275 4756 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.175345 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:30.175329178 +0000 UTC m=+21.325796553 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.175460 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89e5d60f-f690-486f-b078-5ae9f98e1f3a-multus-daemon-config\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.175748 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-etc-kubernetes\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.175782 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89e5d60f-f690-486f-b078-5ae9f98e1f3a-host-run-k8s-cni-cncf-io\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.175957 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89e5d60f-f690-486f-b078-5ae9f98e1f3a-cni-binary-copy\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.176063 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.176128 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.176156 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.176194 4756 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.176161 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f6a6a991-ccf4-409f-bda5-cabc5788ea88-os-release\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.176238 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f6a6a991-ccf4-409f-bda5-cabc5788ea88-cni-binary-copy\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.176264 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:30.176229117 +0000 UTC m=+21.326696492 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.187377 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.190670 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckdq2\" (UniqueName: \"kubernetes.io/projected/89e5d60f-f690-486f-b078-5ae9f98e1f3a-kube-api-access-ckdq2\") pod \"multus-6srkq\" (UID: \"89e5d60f-f690-486f-b078-5ae9f98e1f3a\") " pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.191563 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ld46\" (UniqueName: \"kubernetes.io/projected/f6a6a991-ccf4-409f-bda5-cabc5788ea88-kube-api-access-6ld46\") pod \"multus-additional-cni-plugins-qx7sz\" (UID: \"f6a6a991-ccf4-409f-bda5-cabc5788ea88\") " pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.272348 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.297154 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.304231 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6srkq" Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.309301 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6a6a991_ccf4_409f_bda5_cabc5788ea88.slice/crio-9aa4159b19c4f913b9ebf3a70bf662f79444532753aa38e6336cb7d5fd77c32f WatchSource:0}: Error finding container 9aa4159b19c4f913b9ebf3a70bf662f79444532753aa38e6336cb7d5fd77c32f: Status 404 returned error can't find the container with id 9aa4159b19c4f913b9ebf3a70bf662f79444532753aa38e6336cb7d5fd77c32f Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.318099 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89e5d60f_f690_486f_b078_5ae9f98e1f3a.slice/crio-9d73defa93bb859782084887354d93444bf30891278b312ccd6d50ee3df1e420 WatchSource:0}: Error finding container 9d73defa93bb859782084887354d93444bf30891278b312ccd6d50ee3df1e420: Status 404 returned error can't find the container with id 9d73defa93bb859782084887354d93444bf30891278b312ccd6d50ee3df1e420 Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.348607 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.349767 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gfsqr"] Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.350989 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.354357 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.354593 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.354999 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.355019 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.355262 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.355545 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.357004 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.362650 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376177 4756 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376587 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-script-lib\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376617 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-netns\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376653 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376678 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-kubelet\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376707 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-ovn\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376722 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-config\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376738 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-var-lib-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376754 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-systemd-units\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376768 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-ovn-kubernetes\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376784 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-netd\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376802 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-node-log\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376816 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-bin\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376830 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-etc-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376844 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376867 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovn-node-metrics-cert\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376890 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-slash\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376903 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-log-socket\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376921 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h849g\" (UniqueName: \"kubernetes.io/projected/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-kube-api-access-h849g\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376941 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-systemd\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.376955 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-env-overrides\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377042 4756 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377071 4756 reflector.go:484] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377090 4756 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-config": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377380 4756 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377410 4756 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377429 4756 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377480 4756 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377504 4756 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377534 4756 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377547 4756 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377572 4756 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377578 4756 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377593 4756 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.377643 4756 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": read tcp 38.129.56.121:58250->38.129.56.121:6443: use of closed network connection" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.377655 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-operator/pods/iptables-alerter-4ln5h/status\": read tcp 38.129.56.121:58250->38.129.56.121:6443: use of closed network connection" Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377703 4756 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377740 4756 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377757 4756 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377553 4756 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377765 4756 reflector.go:484] object-"openshift-ovn-kubernetes"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377538 4756 reflector.go:484] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377796 4756 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377803 4756 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377804 4756 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377823 4756 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377830 4756 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377804 4756 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377827 4756 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377867 4756 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377874 4756 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377867 4756 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377868 4756 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.377888 4756 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.386365 4756 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-03 09:05:28 +0000 UTC, rotation deadline is 2026-10-19 11:31:36.828812985 +0000 UTC Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.386436 4756 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6194h21m7.442379988s for next certificate rotation Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.408016 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.423319 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.436051 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.447482 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.458803 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.473723 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477347 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovn-node-metrics-cert\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477392 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-slash\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477411 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-log-socket\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477433 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h849g\" (UniqueName: \"kubernetes.io/projected/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-kube-api-access-h849g\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477535 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-log-socket\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477581 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-systemd\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477650 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-systemd\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477658 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-env-overrides\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477546 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-slash\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.477710 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-script-lib\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.478980 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-netns\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479014 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479036 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-kubelet\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479062 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-ovn\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479077 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-config\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479093 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-var-lib-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479108 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-systemd-units\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479124 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-ovn-kubernetes\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479139 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-netd\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479154 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-node-log\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479170 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-bin\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479188 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-etc-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479203 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479256 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-netns\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479288 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479312 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-kubelet\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.479333 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-ovn\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480603 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-config\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480651 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-var-lib-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480679 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-systemd-units\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480704 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-ovn-kubernetes\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480731 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-netd\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480756 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-node-log\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480779 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-bin\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480804 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-etc-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480814 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-script-lib\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.480831 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-openvswitch\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.481116 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-env-overrides\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.489061 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovn-node-metrics-cert\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.492782 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.507432 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h849g\" (UniqueName: \"kubernetes.io/projected/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-kube-api-access-h849g\") pod \"ovnkube-node-gfsqr\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.509781 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.517098 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.532904 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.546562 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.557964 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.563456 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 12:16:56.819326225 +0000 UTC Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.574695 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.584310 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.602520 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.618051 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.618687 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.619637 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.620570 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.621177 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.621751 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.622328 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.622851 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.623480 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.623985 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.626209 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.626943 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.627847 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.628490 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.629136 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.630906 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.631469 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.632256 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.632923 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.633605 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.635703 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.636345 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.636793 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.637797 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.638218 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.639268 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.639939 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.640856 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.641466 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.643031 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.643550 4756 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.643658 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.647094 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.648577 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.649275 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.649688 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.651492 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.653076 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.653811 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.655041 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.655850 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.657888 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.658492 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.659498 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.660071 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.660991 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.661571 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.662466 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.663160 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.664052 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.664675 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.665681 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.666221 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.666869 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.667891 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.672697 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.674763 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.685431 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: W0203 09:10:29.687848 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3df6b15e_d2d2_4d0e_b7fe_e674fa2a4705.slice/crio-28fa062ddd59b44435884484b416fcd11229a372d50d1998c980b048fdb703eb WatchSource:0}: Error finding container 28fa062ddd59b44435884484b416fcd11229a372d50d1998c980b048fdb703eb: Status 404 returned error can't find the container with id 28fa062ddd59b44435884484b416fcd11229a372d50d1998c980b048fdb703eb Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.711391 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.718817 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d84769a74a7b23c7c9bf08825995e51ea8b11699dcd016c265979fec0d20c85d"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.720106 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cnrvx" event={"ID":"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e","Type":"ContainerStarted","Data":"8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.720161 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cnrvx" event={"ID":"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e","Type":"ContainerStarted","Data":"c264f3ce9e390c8237f0296a973fe43af64256008815f4360a4875c3a56a0bcf"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.722558 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.722605 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.722642 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"7cdd16ad0722165801a8a2cc1a97022b585cef1b1f44d8bc4e0ea9b057e6611a"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.724278 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.724719 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.729701 4756 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945" exitCode=255 Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.729790 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.729867 4756 scope.go:117] "RemoveContainer" containerID="ab6bc4d7606d924bcf2bfa638abe04d8ec4c4186e9cec487bdd13a75317c8afd" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.730057 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.731548 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6srkq" event={"ID":"89e5d60f-f690-486f-b078-5ae9f98e1f3a","Type":"ContainerStarted","Data":"bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.731587 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6srkq" event={"ID":"89e5d60f-f690-486f-b078-5ae9f98e1f3a","Type":"ContainerStarted","Data":"9d73defa93bb859782084887354d93444bf30891278b312ccd6d50ee3df1e420"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.732981 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.733009 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.733020 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"da893d8dbe38df8f4b4fa603c805e05d29a06a049b7133446641afd07277c675"} Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.733333 4756 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-conmon-c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945.scope\": RecentStats: unable to find data in memory cache]" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.734119 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerStarted","Data":"899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.734170 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerStarted","Data":"9aa4159b19c4f913b9ebf3a70bf662f79444532753aa38e6336cb7d5fd77c32f"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.734946 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"28fa062ddd59b44435884484b416fcd11229a372d50d1998c980b048fdb703eb"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.736318 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.736346 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3bbf654be8583549fb230424d7ad4ddc90067917c03560bff6703c1648de5017"} Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.737226 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.763187 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.763400 4756 scope.go:117] "RemoveContainer" containerID="c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945" Feb 03 09:10:29 crc kubenswrapper[4756]: E0203 09:10:29.763584 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.786036 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.831231 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.868631 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.909346 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.952193 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:29 crc kubenswrapper[4756]: I0203 09:10:29.990705 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.025120 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.068473 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.085733 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.085873 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:10:32.085855095 +0000 UTC m=+23.236322470 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.108139 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.149332 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.187137 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.187467 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187369 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.187508 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187520 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.187543 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187549 4756 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187579 4756 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187616 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:32.187599644 +0000 UTC m=+23.338067019 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187619 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187641 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187642 4756 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187651 4756 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187650 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:32.187631795 +0000 UTC m=+23.338099170 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.187384 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187670 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:32.187662656 +0000 UTC m=+23.338130031 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.187708 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:32.187699987 +0000 UTC m=+23.338167472 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.199283 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.253965 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.259054 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.278675 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.319575 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.338868 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.360742 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.364666 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.370147 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.379612 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.399250 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.424254 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.439273 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.478677 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.508221 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.518244 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.559530 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.563919 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 12:20:42.173406406 +0000 UTC Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.594178 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.599386 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.613588 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.613613 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.614039 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.613631 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.614112 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.614291 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.640279 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.666756 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.680349 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.719202 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.739747 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc" exitCode=0 Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.739815 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.741714 4756 generic.go:334] "Generic (PLEG): container finished" podID="f6a6a991-ccf4-409f-bda5-cabc5788ea88" containerID="899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c" exitCode=0 Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.741759 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerDied","Data":"899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c"} Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.744721 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.748290 4756 scope.go:117] "RemoveContainer" containerID="c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945" Feb 03 09:10:30 crc kubenswrapper[4756]: E0203 09:10:30.748537 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.751385 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.760022 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.799064 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.819479 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.839343 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.859980 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.892717 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:30Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.898873 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.918691 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.944018 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.973167 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 03 09:10:30 crc kubenswrapper[4756]: I0203 09:10:30.980884 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.000109 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.018570 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.040140 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.059065 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.078676 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.121571 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.145001 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.185827 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.230989 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab6bc4d7606d924bcf2bfa638abe04d8ec4c4186e9cec487bdd13a75317c8afd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:23Z\\\",\\\"message\\\":\\\"W0203 09:10:12.662870 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0203 09:10:12.663222 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770109812 cert, and key in /tmp/serving-cert-3945495749/serving-signer.crt, /tmp/serving-cert-3945495749/serving-signer.key\\\\nI0203 09:10:13.055546 1 observer_polling.go:159] Starting file observer\\\\nW0203 09:10:13.057894 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0203 09:10:13.058079 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:13.060375 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3945495749/tls.crt::/tmp/serving-cert-3945495749/tls.key\\\\\\\"\\\\nF0203 09:10:23.298718 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.272916 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.309775 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.350380 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.388389 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.427194 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.467618 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.506504 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.545662 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.564733 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 12:55:26.469122699 +0000 UTC Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.588901 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.626015 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.668323 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.707043 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.750732 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.751679 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55"} Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.753431 4756 generic.go:334] "Generic (PLEG): container finished" podID="f6a6a991-ccf4-409f-bda5-cabc5788ea88" containerID="4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33" exitCode=0 Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.753488 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerDied","Data":"4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33"} Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.757284 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.757306 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.757315 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.757324 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.757332 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.757340 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.791783 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.828983 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.868149 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.905888 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.947548 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:31 crc kubenswrapper[4756]: I0203 09:10:31.994484 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:31Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.027626 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.068740 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.102046 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.102225 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:10:36.102211366 +0000 UTC m=+27.252678741 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.109689 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.146489 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.189563 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.202786 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.202832 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.202852 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.202910 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203036 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203053 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203062 4756 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203062 4756 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203122 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:36.203106428 +0000 UTC m=+27.353573813 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203165 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:36.20314216 +0000 UTC m=+27.353609545 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203081 4756 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203219 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:36.203210562 +0000 UTC m=+27.353677947 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203245 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203274 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203293 4756 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.203379 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:36.203348046 +0000 UTC m=+27.353815471 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.225069 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.265743 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.309809 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.565540 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 08:51:39.066827648 +0000 UTC Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.613506 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.613577 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.613665 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.614085 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.614310 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:32 crc kubenswrapper[4756]: E0203 09:10:32.614511 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.763534 4756 generic.go:334] "Generic (PLEG): container finished" podID="f6a6a991-ccf4-409f-bda5-cabc5788ea88" containerID="82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca" exitCode=0 Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.763630 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerDied","Data":"82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca"} Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.785298 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.800367 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.818639 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.833327 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.844660 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.857495 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.867037 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.875156 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.886057 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.899592 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.912399 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.924532 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:32 crc kubenswrapper[4756]: I0203 09:10:32.942960 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:32Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.566121 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 00:27:09.319671025 +0000 UTC Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.772904 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.776677 4756 generic.go:334] "Generic (PLEG): container finished" podID="f6a6a991-ccf4-409f-bda5-cabc5788ea88" containerID="26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73" exitCode=0 Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.776718 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerDied","Data":"26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73"} Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.793758 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.807941 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.828668 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.850347 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.863861 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.876823 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.888129 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.899314 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.908851 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.922725 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.935802 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.947313 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:33 crc kubenswrapper[4756]: I0203 09:10:33.959276 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:33Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.566551 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 05:35:45.276994398 +0000 UTC Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.613891 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.613944 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.613900 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.614023 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.614086 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.614119 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.755154 4756 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.756712 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.756747 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.756758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.756874 4756 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.763081 4756 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.763375 4756 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.764386 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.764411 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.764419 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.764432 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.764461 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:34Z","lastTransitionTime":"2026-02-03T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.776051 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.779267 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.779298 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.779310 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.779326 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.779337 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:34Z","lastTransitionTime":"2026-02-03T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.786785 4756 generic.go:334] "Generic (PLEG): container finished" podID="f6a6a991-ccf4-409f-bda5-cabc5788ea88" containerID="44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd" exitCode=0 Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.786848 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerDied","Data":"44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd"} Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.793319 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.797012 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.797069 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.797086 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.797108 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.797123 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:34Z","lastTransitionTime":"2026-02-03T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.799530 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.813533 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.813593 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.817945 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.817980 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.817992 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.818008 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.818020 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:34Z","lastTransitionTime":"2026-02-03T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.839721 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.876358 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.876378 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.881548 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.881601 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.881616 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.881676 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.881704 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:34Z","lastTransitionTime":"2026-02-03T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.894746 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: E0203 09:10:34.894899 4756 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.896352 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.896391 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.896404 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.896421 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.896433 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:34Z","lastTransitionTime":"2026-02-03T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.899485 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.912557 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.926530 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.939146 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.954008 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.969851 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.980418 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.992535 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:34Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.998758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.998793 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.998802 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.998816 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:34 crc kubenswrapper[4756]: I0203 09:10:34.998826 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:34Z","lastTransitionTime":"2026-02-03T09:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.004349 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.101060 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.101108 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.101120 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.101137 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.101147 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.180830 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.181858 4756 scope.go:117] "RemoveContainer" containerID="c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945" Feb 03 09:10:35 crc kubenswrapper[4756]: E0203 09:10:35.182039 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.203686 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.203743 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.203758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.203775 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.203790 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.306769 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.306847 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.306863 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.306883 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.306898 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.410001 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.410060 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.410074 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.410103 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.410118 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.513493 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.513568 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.513593 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.513619 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.513637 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.567531 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 10:18:37.843682607 +0000 UTC Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.615942 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.615981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.615992 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.616010 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.616035 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.719340 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.719370 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.719378 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.719390 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.719402 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.794498 4756 generic.go:334] "Generic (PLEG): container finished" podID="f6a6a991-ccf4-409f-bda5-cabc5788ea88" containerID="7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf" exitCode=0 Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.794572 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerDied","Data":"7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.802078 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.802373 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.802417 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.811898 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.821870 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.821901 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.821910 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.821923 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.821934 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.825650 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.829103 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.829236 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.843161 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.857064 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.874382 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.889912 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.901406 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.912409 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.921983 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.926686 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.926716 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.926727 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.926744 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.926756 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:35Z","lastTransitionTime":"2026-02-03T09:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.933000 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.945345 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.960175 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.970347 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.983263 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:35 crc kubenswrapper[4756]: I0203 09:10:35.995279 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:35Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.012464 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.025130 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.028359 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.028393 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.028403 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.028418 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.028430 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.038015 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.059126 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.073679 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.084701 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.098100 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.111730 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.121960 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.130695 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.130743 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.130753 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.130775 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.130786 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.134608 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.147570 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.147740 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:10:44.147716074 +0000 UTC m=+35.298183449 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.149669 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.233807 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.233866 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.233885 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.233910 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.233928 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.248402 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.248542 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248590 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248613 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.248611 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248628 4756 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248673 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:44.248657137 +0000 UTC m=+35.399124512 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.248695 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248747 4756 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248761 4756 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248793 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248778 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:44.248769781 +0000 UTC m=+35.399237156 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248829 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248850 4756 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248849 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:44.248831023 +0000 UTC m=+35.399298438 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.248917 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:44.248895265 +0000 UTC m=+35.399362680 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.337068 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.337142 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.337164 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.337191 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.337216 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.438876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.438914 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.438929 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.438945 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.438958 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.541977 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.542026 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.542040 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.542058 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.542074 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.567875 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 03:04:00.71445617 +0000 UTC Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.614110 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.614221 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.614106 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.614279 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.614365 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:36 crc kubenswrapper[4756]: E0203 09:10:36.614507 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.645130 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.645544 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.645712 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.645911 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.646098 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.749412 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.749496 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.749521 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.749547 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.749565 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.808620 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" event={"ID":"f6a6a991-ccf4-409f-bda5-cabc5788ea88","Type":"ContainerStarted","Data":"5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.808688 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.823940 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.843260 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.852109 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.852148 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.852161 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.852179 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.852190 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.857694 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.872848 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.893759 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.911193 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.924781 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.935921 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.945691 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.956027 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.956110 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.956134 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.956166 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.956199 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:36Z","lastTransitionTime":"2026-02-03T09:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.961132 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.971850 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.983539 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:36 crc kubenswrapper[4756]: I0203 09:10:36.995169 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.058588 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.058651 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.058669 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.058691 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.058708 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.161522 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.161880 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.161981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.162083 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.162157 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.264853 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.264912 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.264930 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.264954 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.264974 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.367723 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.367777 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.367794 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.367816 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.367832 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.470424 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.470523 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.470546 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.470578 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.470601 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.568183 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 23:19:44.151849575 +0000 UTC Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.574067 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.574129 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.574154 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.574183 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.574207 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.677044 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.677082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.677090 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.677106 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.677117 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.779376 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.779468 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.779488 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.779511 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.779528 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.813146 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.881797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.881842 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.881851 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.881865 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.881874 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.984136 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.984215 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.984227 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.984243 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:37 crc kubenswrapper[4756]: I0203 09:10:37.984255 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:37Z","lastTransitionTime":"2026-02-03T09:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.086558 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.086594 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.086606 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.086621 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.086632 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.188703 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.188757 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.188774 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.188795 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.188810 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.204937 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.291253 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.291305 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.291319 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.291335 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.291345 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.394868 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.394924 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.394935 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.394959 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.394978 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.498040 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.498131 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.498158 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.498191 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.498215 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.569904 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 03:20:16.505516239 +0000 UTC Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.601379 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.601417 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.601426 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.601439 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.601464 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.613774 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.613816 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:38 crc kubenswrapper[4756]: E0203 09:10:38.613873 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.613778 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:38 crc kubenswrapper[4756]: E0203 09:10:38.613976 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:38 crc kubenswrapper[4756]: E0203 09:10:38.614058 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.703631 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.703681 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.703693 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.703708 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.703720 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.805969 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.806015 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.806023 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.806039 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.806048 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.822772 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/0.log" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.826012 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a" exitCode=1 Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.826067 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.826706 4756 scope.go:117] "RemoveContainer" containerID="b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.839355 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.855342 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.870331 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:38Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:38.416296 6035 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0203 09:10:38.416691 6035 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:38.416758 6035 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:38.416807 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:38.416838 6035 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:38.416857 6035 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:38.416859 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0203 09:10:38.416885 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:38.416896 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:38.416930 6035 factory.go:656] Stopping watch factory\\\\nI0203 09:10:38.416944 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:38.416952 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0203 09:10:38.416959 6035 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.881768 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.893437 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.903334 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.907531 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.907553 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.907563 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.907577 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.907586 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:38Z","lastTransitionTime":"2026-02-03T09:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.913656 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.926849 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.938976 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.957899 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.974231 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:38 crc kubenswrapper[4756]: I0203 09:10:38.986157 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.001343 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:38Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.009672 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.009703 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.009711 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.009724 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.009733 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.112340 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.112397 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.112414 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.112438 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.112489 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.215051 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.215128 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.215160 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.215191 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.215215 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.318054 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.318116 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.318130 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.318146 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.318158 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.420082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.420130 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.420140 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.420156 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.420170 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.522284 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.522319 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.522327 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.522339 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.522347 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.570705 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 01:22:07.915194483 +0000 UTC Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.624069 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.624135 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.624173 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.624202 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.624220 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.628927 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.641664 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.654585 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.665843 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.677435 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.697621 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:38Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:38.416296 6035 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0203 09:10:38.416691 6035 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:38.416758 6035 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:38.416807 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:38.416838 6035 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:38.416857 6035 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:38.416859 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0203 09:10:38.416885 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:38.416896 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:38.416930 6035 factory.go:656] Stopping watch factory\\\\nI0203 09:10:38.416944 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:38.416952 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0203 09:10:38.416959 6035 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.708199 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-626ms"] Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.708553 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.711263 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.711603 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.711603 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.712359 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.716027 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.726682 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.726716 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.726728 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.726744 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.726755 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.729543 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.741026 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.753900 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.763919 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.779530 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.780985 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qq67\" (UniqueName: \"kubernetes.io/projected/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-kube-api-access-2qq67\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.781026 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-host\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.781124 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-serviceca\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.789360 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.800580 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.808855 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.818091 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.828404 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.828456 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.828472 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.828488 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.828500 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.830636 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/1.log" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.831099 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.831235 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/0.log" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.833560 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342" exitCode=1 Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.833604 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.833641 4756 scope.go:117] "RemoveContainer" containerID="b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.834105 4756 scope.go:117] "RemoveContainer" containerID="79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342" Feb 03 09:10:39 crc kubenswrapper[4756]: E0203 09:10:39.834330 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.843554 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.857422 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.872866 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:38Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:38.416296 6035 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0203 09:10:38.416691 6035 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:38.416758 6035 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:38.416807 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:38.416838 6035 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:38.416857 6035 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:38.416859 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0203 09:10:38.416885 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:38.416896 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:38.416930 6035 factory.go:656] Stopping watch factory\\\\nI0203 09:10:38.416944 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:38.416952 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0203 09:10:38.416959 6035 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.882259 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.882291 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qq67\" (UniqueName: \"kubernetes.io/projected/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-kube-api-access-2qq67\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.882414 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-host\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.882492 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-serviceca\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.882546 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-host\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.883468 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-serviceca\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.893130 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.900356 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qq67\" (UniqueName: \"kubernetes.io/projected/34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6-kube-api-access-2qq67\") pod \"node-ca-626ms\" (UID: \"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\") " pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.904519 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.919427 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.931423 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.931486 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.931499 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.931515 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.931528 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:39Z","lastTransitionTime":"2026-02-03T09:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.933287 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.945403 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.957645 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.974478 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:38Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:38.416296 6035 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0203 09:10:38.416691 6035 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:38.416758 6035 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:38.416807 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:38.416838 6035 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:38.416857 6035 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:38.416859 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0203 09:10:38.416885 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:38.416896 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:38.416930 6035 factory.go:656] Stopping watch factory\\\\nI0203 09:10:38.416944 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:38.416952 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0203 09:10:38.416959 6035 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.985114 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:39 crc kubenswrapper[4756]: I0203 09:10:39.996506 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.006401 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.015903 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.022079 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-626ms" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.025190 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.037051 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.037079 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.037089 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.037101 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.037110 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.039348 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.053668 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.066565 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.078310 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.092736 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.107305 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.119270 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.132076 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.139930 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.139983 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.140000 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.140026 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.140045 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.242229 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.242257 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.242267 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.242278 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.242287 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.344764 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.344802 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.344810 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.344822 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.344831 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.449999 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.450081 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.450108 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.450142 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.450166 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.553572 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.553672 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.553708 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.553739 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.553769 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.571781 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 22:14:57.834673762 +0000 UTC Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.613639 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.613702 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.613647 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:40 crc kubenswrapper[4756]: E0203 09:10:40.613853 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:40 crc kubenswrapper[4756]: E0203 09:10:40.613980 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:40 crc kubenswrapper[4756]: E0203 09:10:40.614175 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.656249 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.656305 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.656320 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.656342 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.656357 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.759760 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.759840 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.759859 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.759881 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.759899 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.839220 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-626ms" event={"ID":"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6","Type":"ContainerStarted","Data":"c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.839279 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-626ms" event={"ID":"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6","Type":"ContainerStarted","Data":"4336ea097e875c2a207f54215ad1dc89a9ea55c0f398bffa210c07c8a4fcc9e3"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.842157 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/1.log" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.847951 4756 scope.go:117] "RemoveContainer" containerID="79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342" Feb 03 09:10:40 crc kubenswrapper[4756]: E0203 09:10:40.848135 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.854502 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.868903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.868997 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.869023 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.869056 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.869079 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.870500 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.894571 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5cbc496f6634c89180adc4444f3dc9a3aa43b9d97c0a2ee9067b250965c780a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:38Z\\\",\\\"message\\\":\\\"olicy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:38.416296 6035 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0203 09:10:38.416691 6035 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:38.416758 6035 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:38.416807 6035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:38.416838 6035 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:38.416857 6035 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:38.416859 6035 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0203 09:10:38.416885 6035 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:38.416896 6035 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:38.416930 6035 factory.go:656] Stopping watch factory\\\\nI0203 09:10:38.416944 6035 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:38.416952 6035 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0203 09:10:38.416959 6035 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.909329 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.918745 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.929167 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.940688 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.952988 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.968985 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.971654 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.971715 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.971761 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.971783 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.971797 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:40Z","lastTransitionTime":"2026-02-03T09:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.986958 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:40 crc kubenswrapper[4756]: I0203 09:10:40.999156 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:40Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.012027 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.026232 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.040709 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.059502 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.072725 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.074630 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.074659 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.074668 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.074681 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.074692 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.084720 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.097542 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.117524 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.126411 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.138415 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.149996 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.163330 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.174494 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.177904 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.177952 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.177964 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.177981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.177993 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.188051 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.198525 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.209415 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.219130 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.280053 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.280095 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.280107 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.280126 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.280136 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.382935 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.382999 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.383013 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.383037 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.383051 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.485693 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.486148 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.486342 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.486599 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.486745 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.572850 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 17:12:53.221683078 +0000 UTC Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.589757 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.589816 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.589831 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.589855 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.589873 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.692173 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.692457 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.692533 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.692605 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.692680 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.795627 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.795692 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.795709 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.795734 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.795751 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.899595 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.899656 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.899672 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.899690 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.899723 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:41Z","lastTransitionTime":"2026-02-03T09:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.972898 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl"] Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.973301 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.975722 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.975724 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 03 09:10:41 crc kubenswrapper[4756]: I0203 09:10:41.990517 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.000266 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2vjg\" (UniqueName: \"kubernetes.io/projected/9e9cfd18-1b1c-4c27-a693-22868b869042-kube-api-access-p2vjg\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.000375 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9cfd18-1b1c-4c27-a693-22868b869042-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.000403 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9cfd18-1b1c-4c27-a693-22868b869042-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.000437 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9cfd18-1b1c-4c27-a693-22868b869042-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.000997 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:41Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.002341 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.002391 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.002408 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.002435 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.002474 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.015072 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.031438 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.046640 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.060788 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.073505 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.087888 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.098357 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.100859 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2vjg\" (UniqueName: \"kubernetes.io/projected/9e9cfd18-1b1c-4c27-a693-22868b869042-kube-api-access-p2vjg\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.100939 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9cfd18-1b1c-4c27-a693-22868b869042-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.100972 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9cfd18-1b1c-4c27-a693-22868b869042-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.101008 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9cfd18-1b1c-4c27-a693-22868b869042-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.101599 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e9cfd18-1b1c-4c27-a693-22868b869042-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.101835 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e9cfd18-1b1c-4c27-a693-22868b869042-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.104221 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.104256 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.104267 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.104284 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.104299 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.105766 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e9cfd18-1b1c-4c27-a693-22868b869042-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.109848 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.116068 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2vjg\" (UniqueName: \"kubernetes.io/projected/9e9cfd18-1b1c-4c27-a693-22868b869042-kube-api-access-p2vjg\") pod \"ovnkube-control-plane-749d76644c-lc4vl\" (UID: \"9e9cfd18-1b1c-4c27-a693-22868b869042\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.125051 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.140050 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.152915 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.164251 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.180968 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.206494 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.206648 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.206735 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.206822 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.206916 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.287856 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" Feb 03 09:10:42 crc kubenswrapper[4756]: W0203 09:10:42.303976 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e9cfd18_1b1c_4c27_a693_22868b869042.slice/crio-e39c4ac0a95820fcb1be439fd132f1207cd8de36bd5764ac767e96ec539dd1fc WatchSource:0}: Error finding container e39c4ac0a95820fcb1be439fd132f1207cd8de36bd5764ac767e96ec539dd1fc: Status 404 returned error can't find the container with id e39c4ac0a95820fcb1be439fd132f1207cd8de36bd5764ac767e96ec539dd1fc Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.309061 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.309240 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.309432 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.309869 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.310358 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.414313 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.414355 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.414370 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.414392 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.414406 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.517421 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.517777 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.517788 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.517801 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.517811 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.573762 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 02:27:08.774058671 +0000 UTC Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.613904 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.613908 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:42 crc kubenswrapper[4756]: E0203 09:10:42.614058 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:42 crc kubenswrapper[4756]: E0203 09:10:42.614176 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.613924 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:42 crc kubenswrapper[4756]: E0203 09:10:42.614266 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.620755 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.620784 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.620794 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.620806 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.620815 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.705569 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-k6pzt"] Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.705943 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:42 crc kubenswrapper[4756]: E0203 09:10:42.706035 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.722525 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.722567 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.722576 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.722590 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.722600 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.723942 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.739694 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.759573 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.769397 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.782463 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.795535 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.804867 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.806352 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99zjn\" (UniqueName: \"kubernetes.io/projected/717b7410-dd64-44cb-ba9b-3436d82ebb95-kube-api-access-99zjn\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.806391 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.816923 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.825512 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.825696 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.825783 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.825848 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.825955 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.830411 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.840630 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.849709 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.855375 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" event={"ID":"9e9cfd18-1b1c-4c27-a693-22868b869042","Type":"ContainerStarted","Data":"06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.855636 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" event={"ID":"9e9cfd18-1b1c-4c27-a693-22868b869042","Type":"ContainerStarted","Data":"53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.855734 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" event={"ID":"9e9cfd18-1b1c-4c27-a693-22868b869042","Type":"ContainerStarted","Data":"e39c4ac0a95820fcb1be439fd132f1207cd8de36bd5764ac767e96ec539dd1fc"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.861742 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.872643 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.881055 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.892674 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.902410 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.907058 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99zjn\" (UniqueName: \"kubernetes.io/projected/717b7410-dd64-44cb-ba9b-3436d82ebb95-kube-api-access-99zjn\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.907100 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:42 crc kubenswrapper[4756]: E0203 09:10:42.907229 4756 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:42 crc kubenswrapper[4756]: E0203 09:10:42.907306 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs podName:717b7410-dd64-44cb-ba9b-3436d82ebb95 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:43.407286275 +0000 UTC m=+34.557753650 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs") pod "network-metrics-daemon-k6pzt" (UID: "717b7410-dd64-44cb-ba9b-3436d82ebb95") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.918856 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.925419 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99zjn\" (UniqueName: \"kubernetes.io/projected/717b7410-dd64-44cb-ba9b-3436d82ebb95-kube-api-access-99zjn\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.928205 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.928685 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.928726 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.928736 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.928753 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.928764 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:42Z","lastTransitionTime":"2026-02-03T09:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.940468 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.952678 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.966725 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.978615 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:42 crc kubenswrapper[4756]: I0203 09:10:42.988976 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.000759 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:42Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.013862 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:43Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.027436 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:43Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.031643 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.031674 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.031686 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.031704 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.031717 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.042056 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:43Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.052069 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:43Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.063581 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:43Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.073413 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:43Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.084697 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:43Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.100288 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:43Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.134962 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.135005 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.135019 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.135036 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.135048 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.238370 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.238490 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.238511 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.238541 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.238561 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.340987 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.341034 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.341045 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.341062 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.341084 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.413300 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:43 crc kubenswrapper[4756]: E0203 09:10:43.413606 4756 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:43 crc kubenswrapper[4756]: E0203 09:10:43.413804 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs podName:717b7410-dd64-44cb-ba9b-3436d82ebb95 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:44.413771276 +0000 UTC m=+35.564238691 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs") pod "network-metrics-daemon-k6pzt" (UID: "717b7410-dd64-44cb-ba9b-3436d82ebb95") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.444359 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.444435 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.444503 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.444527 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.444549 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.547930 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.548002 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.548021 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.548049 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.548069 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.574562 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 06:55:53.496091214 +0000 UTC Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.650318 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.650368 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.650379 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.650395 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.650406 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.753824 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.753861 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.753869 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.753883 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.753892 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.858793 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.858833 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.858843 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.858859 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.858870 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.961667 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.961708 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.961717 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.961731 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:43 crc kubenswrapper[4756]: I0203 09:10:43.961740 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:43Z","lastTransitionTime":"2026-02-03T09:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.064217 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.064251 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.064261 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.064276 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.064288 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.167364 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.167423 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.167433 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.167482 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.167495 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.222953 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.223319 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:11:00.223285157 +0000 UTC m=+51.373752582 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.270062 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.270116 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.270125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.270138 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.270147 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.324225 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.324350 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.324404 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.324488 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324565 4756 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324598 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324645 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324663 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:00.324638912 +0000 UTC m=+51.475106297 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324669 4756 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324773 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:00.324742795 +0000 UTC m=+51.475210220 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324672 4756 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324858 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:00.324838878 +0000 UTC m=+51.475306333 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324672 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324902 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.324917 4756 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.325033 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:00.325015844 +0000 UTC m=+51.475483269 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.372619 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.372681 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.372694 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.372715 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.372730 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.425088 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.425238 4756 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.425298 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs podName:717b7410-dd64-44cb-ba9b-3436d82ebb95 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:46.425283635 +0000 UTC m=+37.575751010 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs") pod "network-metrics-daemon-k6pzt" (UID: "717b7410-dd64-44cb-ba9b-3436d82ebb95") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.474864 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.474906 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.474917 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.474933 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.474943 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.575896 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 22:54:42.889601091 +0000 UTC Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.577769 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.577832 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.577847 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.577868 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.577882 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.614136 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.614195 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.614246 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.614155 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.614300 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.614499 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.614588 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.614587 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.680504 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.680547 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.680558 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.680574 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.680585 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.783027 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.783070 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.783081 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.783097 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.783110 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.885482 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.885565 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.885594 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.885622 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.885642 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.980187 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.980228 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.980239 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.980256 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:44 crc kubenswrapper[4756]: I0203 09:10:44.980287 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:44Z","lastTransitionTime":"2026-02-03T09:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:44 crc kubenswrapper[4756]: E0203 09:10:44.998834 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:44Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.002823 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.002858 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.002867 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.002897 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.002908 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: E0203 09:10:45.021520 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:45Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.026742 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.026925 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.026951 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.026981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.027006 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: E0203 09:10:45.046380 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:45Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.050631 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.050667 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.050679 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.050695 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.050708 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: E0203 09:10:45.063795 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:45Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.067603 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.067646 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.067660 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.067676 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.067688 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: E0203 09:10:45.080157 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:45Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:45 crc kubenswrapper[4756]: E0203 09:10:45.080269 4756 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.081607 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.081645 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.081659 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.081681 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.081695 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.184498 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.184561 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.184578 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.184601 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.184623 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.287076 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.287121 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.287132 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.287146 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.287157 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.389642 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.389709 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.389733 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.389761 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.389783 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.492143 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.492226 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.492257 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.492287 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.492307 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.576771 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 23:57:55.081057151 +0000 UTC Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.594724 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.594766 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.594776 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.594788 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.594798 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.696856 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.696901 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.696915 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.696931 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.696948 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.799239 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.799300 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.799315 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.799332 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.799345 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.901827 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.901932 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.901968 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.901996 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:45 crc kubenswrapper[4756]: I0203 09:10:45.902013 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:45Z","lastTransitionTime":"2026-02-03T09:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.003767 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.004080 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.004162 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.004230 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.004309 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.106745 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.106785 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.106801 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.106823 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.106840 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.208673 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.208721 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.208735 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.208751 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.208763 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.310994 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.311042 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.311058 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.311074 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.311085 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.414262 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.414339 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.414363 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.414395 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.414416 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.446628 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:46 crc kubenswrapper[4756]: E0203 09:10:46.446971 4756 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:46 crc kubenswrapper[4756]: E0203 09:10:46.447111 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs podName:717b7410-dd64-44cb-ba9b-3436d82ebb95 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:50.447075106 +0000 UTC m=+41.597542571 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs") pod "network-metrics-daemon-k6pzt" (UID: "717b7410-dd64-44cb-ba9b-3436d82ebb95") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.517166 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.517221 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.517236 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.517256 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.517271 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.577882 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 06:18:49.319170978 +0000 UTC Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.614146 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.614255 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.614146 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:46 crc kubenswrapper[4756]: E0203 09:10:46.614521 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.614150 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:46 crc kubenswrapper[4756]: E0203 09:10:46.614605 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:46 crc kubenswrapper[4756]: E0203 09:10:46.614277 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:46 crc kubenswrapper[4756]: E0203 09:10:46.614656 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.620777 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.620831 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.620851 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.620875 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.620894 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.723395 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.723672 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.723767 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.723852 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.723920 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.827134 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.827173 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.827184 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.827200 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.827218 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.929262 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.929296 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.929305 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.929319 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:46 crc kubenswrapper[4756]: I0203 09:10:46.929328 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:46Z","lastTransitionTime":"2026-02-03T09:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.031408 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.031501 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.031511 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.031526 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.031535 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.133649 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.133689 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.133700 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.133715 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.133726 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.236431 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.236496 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.236506 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.236522 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.236533 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.339751 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.339835 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.339850 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.339869 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.339885 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.442356 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.442400 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.442410 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.442422 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.442432 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.546574 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.546643 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.546661 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.546685 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.546703 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.579060 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 19:08:00.603735123 +0000 UTC Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.648674 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.648711 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.648722 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.648737 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.648748 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.751003 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.751086 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.751107 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.751164 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.751184 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.854181 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.854243 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.854262 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.854290 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.854317 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.957635 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.957680 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.957708 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.957723 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:47 crc kubenswrapper[4756]: I0203 09:10:47.957733 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:47Z","lastTransitionTime":"2026-02-03T09:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.060183 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.060244 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.060261 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.060284 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.060302 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.163551 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.163597 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.163610 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.163664 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.163676 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.266301 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.266368 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.266385 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.266408 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.266510 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.369026 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.369100 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.369122 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.369679 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.369732 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.471680 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.471738 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.471753 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.471774 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.471790 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.574803 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.574865 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.574880 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.574897 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.574911 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.579379 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 09:57:08.629058154 +0000 UTC Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.613769 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.613929 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.613930 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.614108 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:48 crc kubenswrapper[4756]: E0203 09:10:48.614098 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:48 crc kubenswrapper[4756]: E0203 09:10:48.614279 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:48 crc kubenswrapper[4756]: E0203 09:10:48.614392 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:48 crc kubenswrapper[4756]: E0203 09:10:48.614690 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.678826 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.678882 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.678900 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.678923 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.678940 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.780772 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.780815 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.780826 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.780845 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.780858 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.883139 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.883186 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.883199 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.883213 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.883225 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.985870 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.985925 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.985942 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.985958 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:48 crc kubenswrapper[4756]: I0203 09:10:48.985968 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:48Z","lastTransitionTime":"2026-02-03T09:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.088229 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.088260 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.088270 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.088282 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.088291 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.191517 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.191572 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.191585 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.191606 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.191619 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.295214 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.295648 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.295737 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.295840 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.295935 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.399618 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.399666 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.399682 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.399705 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.399719 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.503670 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.504027 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.504198 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.504355 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.504766 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.580171 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 19:13:18.552538712 +0000 UTC Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.607765 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.607804 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.607815 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.607834 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.607846 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.626248 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.641348 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.667642 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.679507 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.691553 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.709234 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.709757 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.709797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.709809 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.709827 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.709839 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.724596 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.736783 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.747649 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.760611 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.772211 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.781623 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.796637 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.807435 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.812315 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.812387 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.812397 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.812411 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.812420 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.818810 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.830424 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:49Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.915499 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.915538 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.915546 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.915560 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:49 crc kubenswrapper[4756]: I0203 09:10:49.915568 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:49Z","lastTransitionTime":"2026-02-03T09:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.018075 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.018121 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.018136 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.018158 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.018172 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.120667 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.120714 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.120729 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.120749 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.120764 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.222836 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.222874 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.222885 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.222897 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.222906 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.325206 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.325264 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.325278 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.325298 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.325313 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.428331 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.428402 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.428424 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.428491 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.428524 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.489322 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:50 crc kubenswrapper[4756]: E0203 09:10:50.489478 4756 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:50 crc kubenswrapper[4756]: E0203 09:10:50.489543 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs podName:717b7410-dd64-44cb-ba9b-3436d82ebb95 nodeName:}" failed. No retries permitted until 2026-02-03 09:10:58.489523321 +0000 UTC m=+49.639990696 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs") pod "network-metrics-daemon-k6pzt" (UID: "717b7410-dd64-44cb-ba9b-3436d82ebb95") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.530244 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.530287 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.530302 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.530352 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.530362 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.581934 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 04:55:47.549776056 +0000 UTC Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.613706 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.613710 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.613723 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.613731 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:50 crc kubenswrapper[4756]: E0203 09:10:50.614013 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:50 crc kubenswrapper[4756]: E0203 09:10:50.614177 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.614261 4756 scope.go:117] "RemoveContainer" containerID="c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945" Feb 03 09:10:50 crc kubenswrapper[4756]: E0203 09:10:50.614383 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:50 crc kubenswrapper[4756]: E0203 09:10:50.614487 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.633805 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.633839 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.633853 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.633868 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.633879 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.736461 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.736496 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.736506 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.736521 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.736530 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.838778 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.838826 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.838841 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.838860 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.838875 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.884117 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.885618 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.885941 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.897980 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:50Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.908332 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:50Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.921730 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:50Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.940869 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:50Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.941375 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.941534 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.941625 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.941713 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.941792 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:50Z","lastTransitionTime":"2026-02-03T09:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.953197 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:50Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.971322 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:50Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.982955 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:50Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:50 crc kubenswrapper[4756]: I0203 09:10:50.994963 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:50Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.007317 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:51Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.021701 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:51Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.033907 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:51Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.044030 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.044060 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.044068 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.044081 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.044090 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.044846 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:51Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.055163 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:51Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.065279 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:51Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.077286 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:51Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.090030 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:51Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.145948 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.145994 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.146007 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.146065 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.146079 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.248752 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.248797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.248811 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.248827 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.248838 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.351396 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.351437 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.351577 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.351593 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.351602 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.454666 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.454900 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.454919 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.455386 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.455731 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.558781 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.558827 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.558839 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.558854 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.558864 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.582697 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 20:16:11.903179371 +0000 UTC Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.662108 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.662159 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.662173 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.662191 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.662203 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.764739 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.764805 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.764816 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.764829 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.764838 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.867299 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.867635 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.867747 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.867860 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.867951 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.970876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.970913 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.970921 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.970936 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:51 crc kubenswrapper[4756]: I0203 09:10:51.970946 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:51Z","lastTransitionTime":"2026-02-03T09:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.072949 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.072996 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.073007 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.073024 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.073037 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.175761 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.175803 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.175815 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.175830 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.175841 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.278899 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.278981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.278999 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.279047 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.279065 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.381050 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.381089 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.381098 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.381113 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.381121 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.484370 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.484416 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.484428 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.484462 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.484472 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.582899 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 23:44:46.821088866 +0000 UTC Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.587064 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.587105 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.587119 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.587138 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.587152 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.613807 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.613839 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:52 crc kubenswrapper[4756]: E0203 09:10:52.613915 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.613985 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:52 crc kubenswrapper[4756]: E0203 09:10:52.614123 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:52 crc kubenswrapper[4756]: E0203 09:10:52.614190 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.614373 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:52 crc kubenswrapper[4756]: E0203 09:10:52.614585 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.689981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.690029 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.690042 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.690058 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.690071 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.792366 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.792636 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.792723 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.792804 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.792888 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.898211 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.898254 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.898266 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.898282 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:52 crc kubenswrapper[4756]: I0203 09:10:52.898296 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:52Z","lastTransitionTime":"2026-02-03T09:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.001854 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.001909 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.001917 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.001932 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.001941 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.105145 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.105200 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.105209 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.105226 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.105236 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.208300 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.208340 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.208348 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.208383 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.208391 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.310554 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.310621 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.310632 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.310669 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.310684 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.413779 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.413856 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.413878 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.413907 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.413928 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.516424 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.516511 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.516527 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.516547 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.516563 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.583677 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 16:26:27.573659193 +0000 UTC Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.618981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.619016 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.619024 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.619037 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.619045 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.721950 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.721989 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.721997 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.722009 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.722018 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.824241 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.824276 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.824287 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.824302 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.824313 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.926927 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.926972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.926985 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.927001 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:53 crc kubenswrapper[4756]: I0203 09:10:53.927012 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:53Z","lastTransitionTime":"2026-02-03T09:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.030823 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.030903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.030914 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.030930 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.030943 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.134063 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.134105 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.134115 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.134130 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.134140 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.236430 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.236489 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.236501 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.236517 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.236527 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.339401 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.339520 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.339547 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.339577 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.339599 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.442804 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.442866 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.442898 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.442930 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.442956 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.545460 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.545502 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.545512 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.545529 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.545541 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.584355 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 20:30:16.035259021 +0000 UTC Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.613695 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.613700 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:54 crc kubenswrapper[4756]: E0203 09:10:54.613922 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.613701 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.613701 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:54 crc kubenswrapper[4756]: E0203 09:10:54.614067 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:54 crc kubenswrapper[4756]: E0203 09:10:54.614205 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:54 crc kubenswrapper[4756]: E0203 09:10:54.614283 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.647851 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.647911 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.647924 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.647938 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.647946 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.750613 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.750667 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.750684 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.750706 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.750723 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.853741 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.853787 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.853796 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.853812 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.853831 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.956812 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.956853 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.956864 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.956880 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:54 crc kubenswrapper[4756]: I0203 09:10:54.956892 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:54Z","lastTransitionTime":"2026-02-03T09:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.060088 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.060378 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.060477 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.060602 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.060686 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.129826 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.129856 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.129864 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.129876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.129885 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: E0203 09:10:55.143067 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.146591 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.146627 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.146637 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.146650 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.146661 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: E0203 09:10:55.160797 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.163907 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.163945 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.163958 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.163974 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.163986 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: E0203 09:10:55.174370 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.177808 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.177838 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.177848 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.177865 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.177874 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: E0203 09:10:55.188299 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.192079 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.192157 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.192166 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.192180 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.192190 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: E0203 09:10:55.207286 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:55 crc kubenswrapper[4756]: E0203 09:10:55.207399 4756 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.208925 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.208967 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.208978 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.208991 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.209000 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.311699 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.311759 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.311771 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.311788 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.311800 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.414704 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.414742 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.414754 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.414772 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.414784 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.517842 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.517886 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.517900 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.517921 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.517934 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.585506 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 10:31:03.131039236 +0000 UTC Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.614988 4756 scope.go:117] "RemoveContainer" containerID="79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.619571 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.619603 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.619615 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.619629 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.619640 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.722847 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.722928 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.722953 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.722983 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.723005 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.825479 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.825537 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.825555 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.825579 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.825596 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.901906 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/1.log" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.904671 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.905026 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.918170 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.927988 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.928027 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.928040 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.928055 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.928066 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:55Z","lastTransitionTime":"2026-02-03T09:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.954226 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.979867 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:55 crc kubenswrapper[4756]: I0203 09:10:55.994944 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:55Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.018523 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.030226 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.030273 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.030285 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.030303 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.030315 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.034558 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.045504 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.064281 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.089396 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.103288 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.115114 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.125203 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.132653 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.132689 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.132697 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.132709 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.132718 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.135754 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.146377 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.159838 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.169966 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.234677 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.234708 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.234716 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.234729 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.234737 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.338047 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.338125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.338149 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.338184 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.338231 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.486269 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.486593 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.486602 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.486617 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.486625 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.586046 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 22:08:37.515937181 +0000 UTC Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.589322 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.589364 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.589375 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.589392 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.589405 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.613773 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.613821 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:56 crc kubenswrapper[4756]: E0203 09:10:56.613901 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.613915 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:56 crc kubenswrapper[4756]: E0203 09:10:56.614032 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.613777 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:56 crc kubenswrapper[4756]: E0203 09:10:56.614127 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:56 crc kubenswrapper[4756]: E0203 09:10:56.614179 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.692532 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.692609 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.692632 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.692663 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.692687 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.794930 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.795357 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.795665 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.795843 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.796047 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.898903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.898945 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.898957 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.898974 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.898988 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:56Z","lastTransitionTime":"2026-02-03T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.910046 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/2.log" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.910909 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/1.log" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.913862 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138" exitCode=1 Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.913919 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138"} Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.913986 4756 scope.go:117] "RemoveContainer" containerID="79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.914814 4756 scope.go:117] "RemoveContainer" containerID="9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138" Feb 03 09:10:56 crc kubenswrapper[4756]: E0203 09:10:56.915032 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.927644 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.941696 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.960646 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.978126 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.988411 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:56 crc kubenswrapper[4756]: I0203 09:10:56.998401 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:56Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.000672 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.000701 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.000710 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.000723 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.000732 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.015119 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b470f6ee22cf39c043cd253d272081798bdc7d7542291d5d1c29ed2c141342\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"message\\\":\\\"/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697772 6164 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.697827 6164 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.697995 6164 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0203 09:10:39.698193 6164 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698474 6164 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:39.698499 6164 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.698602 6164 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0203 09:10:39.699156 6164 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.025307 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.037430 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.048200 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.059772 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.072997 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.083079 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.092379 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.102290 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.102535 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.102578 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.102588 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.102604 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.102614 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.115214 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.204215 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.204253 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.204261 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.204275 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.204284 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.306864 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.306923 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.306936 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.306952 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.306964 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.410458 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.410500 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.410512 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.410527 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.410537 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.514490 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.514564 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.514585 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.514608 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.514622 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.586529 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 22:07:40.618081475 +0000 UTC Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.616873 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.616917 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.616926 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.616938 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.616947 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.719485 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.719526 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.719537 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.719551 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.719562 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.822431 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.822507 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.822516 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.822531 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.822543 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.923420 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/2.log" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.924505 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.924573 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.924590 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.924612 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.924627 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:57Z","lastTransitionTime":"2026-02-03T09:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.928412 4756 scope.go:117] "RemoveContainer" containerID="9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138" Feb 03 09:10:57 crc kubenswrapper[4756]: E0203 09:10:57.928652 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.942555 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.959271 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:57 crc kubenswrapper[4756]: I0203 09:10:57.981154 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:57Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.007084 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.022206 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.027673 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.027739 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.027766 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.027799 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.027823 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.041790 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.056636 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.078162 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.095166 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.111547 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.130122 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.131544 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.131602 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.131621 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.131674 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.131700 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.151206 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.170458 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.187747 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.202884 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.216888 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.233760 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.233808 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.233820 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.233840 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.233853 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.337296 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.337672 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.337756 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.337900 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.337981 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.440940 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.441009 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.441023 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.441044 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.441057 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.505491 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:58 crc kubenswrapper[4756]: E0203 09:10:58.505733 4756 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:58 crc kubenswrapper[4756]: E0203 09:10:58.505870 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs podName:717b7410-dd64-44cb-ba9b-3436d82ebb95 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:14.505842061 +0000 UTC m=+65.656309436 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs") pod "network-metrics-daemon-k6pzt" (UID: "717b7410-dd64-44cb-ba9b-3436d82ebb95") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.544924 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.545335 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.545627 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.545800 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.546006 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.587465 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 17:40:02.584143542 +0000 UTC Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.613961 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.614402 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.614465 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.614680 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:10:58 crc kubenswrapper[4756]: E0203 09:10:58.614899 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:10:58 crc kubenswrapper[4756]: E0203 09:10:58.615017 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:10:58 crc kubenswrapper[4756]: E0203 09:10:58.615197 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:10:58 crc kubenswrapper[4756]: E0203 09:10:58.615346 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.652475 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.652541 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.652555 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.652571 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.652585 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.697824 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.710196 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.713400 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.729830 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.751813 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.754995 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.755025 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.755035 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.755050 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.755062 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.766757 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.777982 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.788680 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.805499 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.819714 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.832212 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.843664 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.855057 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.858074 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.858115 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.858128 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.858146 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.858157 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.869651 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.881901 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.893984 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.905994 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.920498 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:58Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.960713 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.960762 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.960774 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.960791 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:58 crc kubenswrapper[4756]: I0203 09:10:58.960803 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:58Z","lastTransitionTime":"2026-02-03T09:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.063312 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.063385 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.063404 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.063422 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.063434 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.165239 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.165270 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.165280 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.165306 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.165318 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.267490 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.267527 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.267538 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.267550 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.267559 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.369833 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.369876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.369888 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.369903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.369914 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.472976 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.473536 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.473753 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.474204 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.474588 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.577131 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.577421 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.577530 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.577635 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.577752 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.589390 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 23:01:25.743908033 +0000 UTC Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.628827 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.642131 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.655939 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.672894 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.680949 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.681368 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.681519 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.681678 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.681831 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.690370 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.704570 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.715155 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.723274 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.739507 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.748948 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.760763 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.773364 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.784037 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.784822 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.784870 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.784880 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.784895 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.784908 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.796406 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.807047 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.826548 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.836174 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:10:59Z is after 2025-08-24T17:21:41Z" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.887412 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.887483 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.887504 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.887522 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.887534 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.989142 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.989202 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.989213 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.989229 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:10:59 crc kubenswrapper[4756]: I0203 09:10:59.989240 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:10:59Z","lastTransitionTime":"2026-02-03T09:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.091283 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.091332 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.091340 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.091354 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.091373 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.193661 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.193732 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.193752 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.193776 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.193837 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.228315 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.228529 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:11:32.228496139 +0000 UTC m=+83.378963544 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.296975 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.297063 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.297100 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.297136 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.297159 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.329932 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.330019 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.330068 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.330104 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330122 4756 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330141 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330174 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330182 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330194 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330201 4756 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330205 4756 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330209 4756 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330203 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:32.330180334 +0000 UTC m=+83.480647719 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330275 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:32.330260817 +0000 UTC m=+83.480728202 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330296 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:32.330284468 +0000 UTC m=+83.480751853 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.330315 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:32.330305018 +0000 UTC m=+83.480772403 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.400184 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.400230 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.400244 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.400263 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.400277 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.503103 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.503154 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.503164 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.503182 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.503192 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.589842 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 05:00:45.148356061 +0000 UTC Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.606065 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.606139 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.606165 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.606213 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.606238 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.613436 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.613509 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.613630 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.613691 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.613756 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.613467 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.613827 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:00 crc kubenswrapper[4756]: E0203 09:11:00.613974 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.708784 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.708828 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.708839 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.708857 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.708871 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.811889 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.811958 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.811975 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.811995 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.812010 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.914072 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.914114 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.914125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.914140 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:00 crc kubenswrapper[4756]: I0203 09:11:00.914151 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:00Z","lastTransitionTime":"2026-02-03T09:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.016239 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.016288 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.016299 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.016314 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.016326 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.118963 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.119279 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.119316 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.119338 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.119351 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.221619 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.221665 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.221678 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.221696 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.221708 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.326330 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.326607 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.326708 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.326822 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.326893 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.429110 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.429351 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.429465 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.429562 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.429675 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.532117 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.532180 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.532193 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.532208 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.532220 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.590620 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 14:21:49.756309226 +0000 UTC Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.635329 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.635715 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.635840 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.635969 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.636073 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.738009 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.738281 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.738377 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.738550 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.738705 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.841599 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.841651 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.841664 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.841679 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.841688 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.944029 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.944082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.944100 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.944120 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:01 crc kubenswrapper[4756]: I0203 09:11:01.944136 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:01Z","lastTransitionTime":"2026-02-03T09:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.047647 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.047925 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.048115 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.048202 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.048276 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.151169 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.151248 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.151271 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.151297 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.151315 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.254395 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.254455 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.254468 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.254485 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.254499 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.358234 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.358315 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.358332 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.358357 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.358373 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.460772 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.460837 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.460856 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.460886 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.460911 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.564023 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.564060 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.564070 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.564085 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.564095 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.590833 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 09:05:11.161674442 +0000 UTC Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.613198 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.613249 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.613204 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.613260 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:02 crc kubenswrapper[4756]: E0203 09:11:02.613432 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:02 crc kubenswrapper[4756]: E0203 09:11:02.613584 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:02 crc kubenswrapper[4756]: E0203 09:11:02.613800 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:02 crc kubenswrapper[4756]: E0203 09:11:02.613899 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.667679 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.667738 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.667758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.667781 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.667798 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.770855 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.770919 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.770934 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.770955 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.770967 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.873513 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.873571 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.873594 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.873614 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.873628 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.976082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.976145 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.976172 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.976201 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:02 crc kubenswrapper[4756]: I0203 09:11:02.976223 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:02Z","lastTransitionTime":"2026-02-03T09:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.079647 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.079693 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.079711 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.079732 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.079747 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.182272 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.182322 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.182332 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.182351 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.182362 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.284647 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.284717 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.284735 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.284760 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.284777 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.387462 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.387503 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.387530 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.387546 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.387555 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.490944 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.491210 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.491274 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.491340 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.491400 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.591206 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 18:38:13.236567275 +0000 UTC Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.594227 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.594277 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.594291 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.594309 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.594324 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.697953 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.698751 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.698972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.699135 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.699275 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.802284 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.802332 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.802345 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.802367 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.802383 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.905566 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.905596 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.905604 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.905617 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:03 crc kubenswrapper[4756]: I0203 09:11:03.905627 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:03Z","lastTransitionTime":"2026-02-03T09:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.007429 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.007476 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.007484 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.007498 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.007509 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.110322 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.110483 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.110523 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.110549 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.110565 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.213139 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.213255 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.213268 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.213281 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.213291 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.316204 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.316263 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.316274 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.316291 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.316303 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.418711 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.418756 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.418765 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.418779 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.418790 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.522311 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.522389 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.522406 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.522431 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.522483 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.591648 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 22:26:11.099419448 +0000 UTC Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.613816 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.613872 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.613877 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:04 crc kubenswrapper[4756]: E0203 09:11:04.613956 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.613979 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:04 crc kubenswrapper[4756]: E0203 09:11:04.614031 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:04 crc kubenswrapper[4756]: E0203 09:11:04.614171 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:04 crc kubenswrapper[4756]: E0203 09:11:04.614388 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.625402 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.625471 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.625482 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.625499 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.625511 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.727815 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.727848 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.727856 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.727869 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.727879 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.830651 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.830724 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.830743 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.830768 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.830789 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.934025 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.934089 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.934103 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.934127 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:04 crc kubenswrapper[4756]: I0203 09:11:04.934148 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:04Z","lastTransitionTime":"2026-02-03T09:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.037086 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.037138 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.037156 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.037177 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.037192 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.140853 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.140928 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.140946 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.140972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.140991 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.244053 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.244112 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.244125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.244142 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.244154 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.346975 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.347040 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.347058 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.347082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.347100 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.450071 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.450124 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.450137 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.450159 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.450174 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.481583 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.481635 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.481644 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.481660 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.481673 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: E0203 09:11:05.492548 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:05Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.495964 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.495995 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.496004 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.496017 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.496026 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: E0203 09:11:05.509771 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:05Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.513853 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.513898 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.513913 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.513934 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.513946 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: E0203 09:11:05.527232 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:05Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.531791 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.531851 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.531875 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.531906 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.531930 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: E0203 09:11:05.554245 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:05Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.559734 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.559783 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.559800 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.559824 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.559842 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: E0203 09:11:05.575909 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:05Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:05 crc kubenswrapper[4756]: E0203 09:11:05.576077 4756 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.577701 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.577736 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.577749 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.577768 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.577783 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.592622 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 11:09:07.552811579 +0000 UTC Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.680553 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.680867 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.680972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.681070 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.681308 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.783421 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.783734 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.783770 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.783799 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.783865 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.886589 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.886628 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.886637 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.886651 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.886661 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.991199 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.991249 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.991265 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.991283 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:05 crc kubenswrapper[4756]: I0203 09:11:05.991296 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:05Z","lastTransitionTime":"2026-02-03T09:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.094083 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.094157 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.094170 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.094186 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.094198 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.196896 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.196964 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.196982 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.197006 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.197027 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.299631 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.299682 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.299695 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.299713 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.299728 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.402785 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.402838 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.402846 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.402861 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.402891 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.506091 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.506182 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.506205 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.506236 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.506263 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.593510 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 00:59:04.282930091 +0000 UTC Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.609053 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.609138 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.609168 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.609204 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.609232 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.613307 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.613354 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.613363 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.613423 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:06 crc kubenswrapper[4756]: E0203 09:11:06.613557 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:06 crc kubenswrapper[4756]: E0203 09:11:06.613738 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:06 crc kubenswrapper[4756]: E0203 09:11:06.613772 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:06 crc kubenswrapper[4756]: E0203 09:11:06.613811 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.711544 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.711587 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.711598 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.711613 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.711622 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.814024 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.814072 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.814080 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.814093 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.814103 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.916384 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.916430 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.916477 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.916500 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:06 crc kubenswrapper[4756]: I0203 09:11:06.916512 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:06Z","lastTransitionTime":"2026-02-03T09:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.019542 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.019602 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.019618 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.019642 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.019659 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.121639 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.121691 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.121701 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.121717 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.121728 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.224020 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.224062 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.224069 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.224082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.224090 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.326289 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.326328 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.326336 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.326351 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.326367 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.428913 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.428966 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.428977 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.428994 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.429005 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.531758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.531807 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.531823 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.531841 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.531852 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.594514 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 15:42:39.821111719 +0000 UTC Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.634661 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.634732 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.634755 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.634784 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.634808 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.736950 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.736996 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.737005 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.737019 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.737028 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.839263 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.839566 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.839594 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.839615 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.839639 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.943726 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.943781 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.943797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.943817 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:07 crc kubenswrapper[4756]: I0203 09:11:07.943831 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:07Z","lastTransitionTime":"2026-02-03T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.047203 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.047307 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.047334 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.047365 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.047394 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.149953 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.150033 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.150108 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.150140 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.150165 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.253090 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.253146 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.253156 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.253177 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.253189 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.356192 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.356269 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.356293 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.356322 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.356344 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.459488 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.459539 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.459554 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.459574 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.459591 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.563289 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.563323 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.563332 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.563344 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.563354 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.595055 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 15:42:14.531719408 +0000 UTC Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.613392 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.613421 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.613524 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.613521 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:08 crc kubenswrapper[4756]: E0203 09:11:08.613591 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:08 crc kubenswrapper[4756]: E0203 09:11:08.613688 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:08 crc kubenswrapper[4756]: E0203 09:11:08.613793 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:08 crc kubenswrapper[4756]: E0203 09:11:08.613901 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.665774 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.665846 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.665862 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.665885 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.665899 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.768483 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.768515 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.768523 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.768539 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.768549 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.870250 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.870277 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.870285 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.870297 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.870305 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.972355 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.972412 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.972429 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.972487 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:08 crc kubenswrapper[4756]: I0203 09:11:08.972504 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:08Z","lastTransitionTime":"2026-02-03T09:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.075187 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.075232 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.075248 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.075270 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.075283 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.177790 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.177832 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.177844 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.177859 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.177871 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.279261 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.280431 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.280538 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.280567 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.280594 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.280614 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.294483 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.310822 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.330176 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.348463 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.360730 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.371884 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.382683 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.382720 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.382729 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.382741 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.382749 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.388012 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.406163 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.417190 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.429063 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.443061 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.456353 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.468499 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.482047 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.485591 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.485661 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.485679 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.486074 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.486119 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.497977 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.511021 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.530763 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.588286 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.588341 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.588353 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.588369 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.588382 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.595161 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 17:48:27.812229474 +0000 UTC Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.626340 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.637658 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.654652 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.664560 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.674200 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.685706 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.690615 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.690651 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.690659 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.690673 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.690683 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.702269 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.712232 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.722883 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.732925 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.746396 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.756550 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.769781 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.788587 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.792619 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.792651 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.792659 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.792673 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.792681 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.799832 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.816146 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.830161 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:09Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.894690 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.894732 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.894744 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.894759 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.894772 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.997950 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.998275 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.998358 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.998473 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:09 crc kubenswrapper[4756]: I0203 09:11:09.998572 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:09Z","lastTransitionTime":"2026-02-03T09:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.101203 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.101240 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.101250 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.101263 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.101272 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.204158 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.204216 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.204229 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.204254 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.204265 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.306748 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.306979 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.307039 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.307125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.307189 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.409422 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.409500 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.409512 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.409529 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.409540 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.512225 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.512294 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.512318 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.512345 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.512366 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.595324 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 09:31:51.180827993 +0000 UTC Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.613396 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.613473 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:10 crc kubenswrapper[4756]: E0203 09:11:10.613537 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.613409 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.613696 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:10 crc kubenswrapper[4756]: E0203 09:11:10.613776 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:10 crc kubenswrapper[4756]: E0203 09:11:10.613893 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:10 crc kubenswrapper[4756]: E0203 09:11:10.614049 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.614235 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.614302 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.614324 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.614354 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.614379 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.716406 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.716467 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.716479 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.716493 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.716504 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.818907 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.819054 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.819067 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.819084 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.819437 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.922389 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.922490 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.922515 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.922549 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:10 crc kubenswrapper[4756]: I0203 09:11:10.922570 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:10Z","lastTransitionTime":"2026-02-03T09:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.025130 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.025164 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.025172 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.025188 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.025198 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.127365 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.127401 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.127410 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.127422 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.127430 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.229851 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.229891 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.229902 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.229918 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.229929 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.332721 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.333083 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.333208 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.333312 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.333385 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.436283 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.436320 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.436329 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.436350 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.436360 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.539415 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.539937 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.540015 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.540084 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.540174 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.596376 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 18:13:42.882084629 +0000 UTC Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.614605 4756 scope.go:117] "RemoveContainer" containerID="9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138" Feb 03 09:11:11 crc kubenswrapper[4756]: E0203 09:11:11.614803 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.643052 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.643322 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.643418 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.643526 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.643594 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.746800 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.746835 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.746845 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.746860 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.746873 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.849015 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.849052 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.849063 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.849079 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.849090 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.951249 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.951289 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.951301 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.951315 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:11 crc kubenswrapper[4756]: I0203 09:11:11.951326 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:11Z","lastTransitionTime":"2026-02-03T09:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.053180 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.053232 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.053242 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.053255 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.053265 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.155559 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.155614 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.155625 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.155643 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.155655 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.257951 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.258272 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.258349 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.258463 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.258550 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.360545 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.360594 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.360612 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.360633 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.360644 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.463338 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.463380 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.463389 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.463402 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.463413 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.566209 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.566261 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.566276 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.566292 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.566304 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.597597 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 15:45:03.984164799 +0000 UTC Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.613975 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.614067 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:12 crc kubenswrapper[4756]: E0203 09:11:12.614103 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.613997 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.614085 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:12 crc kubenswrapper[4756]: E0203 09:11:12.614412 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:12 crc kubenswrapper[4756]: E0203 09:11:12.614583 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:12 crc kubenswrapper[4756]: E0203 09:11:12.614710 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.668582 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.668655 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.668679 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.668707 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.668731 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.771269 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.771299 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.771310 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.771324 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.771335 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.874137 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.874186 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.874203 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.874225 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.874241 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.976546 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.976596 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.976608 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.976625 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:12 crc kubenswrapper[4756]: I0203 09:11:12.976637 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:12Z","lastTransitionTime":"2026-02-03T09:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.078938 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.079003 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.079026 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.079055 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.079071 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.181313 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.181358 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.181370 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.181389 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.181402 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.284494 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.284554 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.284571 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.284596 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.284613 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.386843 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.387121 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.387221 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.387312 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.387419 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.489179 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.489375 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.489497 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.489811 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.489898 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.593095 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.593131 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.593141 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.593154 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.593163 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.598464 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 05:07:59.955143819 +0000 UTC Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.695528 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.695561 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.695570 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.695582 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.695591 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.798244 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.798275 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.798286 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.798302 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.798314 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.900192 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.900512 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.900698 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.900840 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:13 crc kubenswrapper[4756]: I0203 09:11:13.901025 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:13Z","lastTransitionTime":"2026-02-03T09:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.003780 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.003834 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.003845 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.003862 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.003876 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.106297 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.106339 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.106348 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.106362 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.106372 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.209069 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.209125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.209135 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.209156 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.209169 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.311935 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.311982 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.311991 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.312007 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.312017 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.414933 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.415009 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.415020 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.415038 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.415051 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.517483 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.517531 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.517544 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.517562 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.517576 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.579827 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:14 crc kubenswrapper[4756]: E0203 09:11:14.580010 4756 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:11:14 crc kubenswrapper[4756]: E0203 09:11:14.580085 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs podName:717b7410-dd64-44cb-ba9b-3436d82ebb95 nodeName:}" failed. No retries permitted until 2026-02-03 09:11:46.580068956 +0000 UTC m=+97.730536331 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs") pod "network-metrics-daemon-k6pzt" (UID: "717b7410-dd64-44cb-ba9b-3436d82ebb95") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.598964 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 18:06:43.712016522 +0000 UTC Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.613126 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.613199 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.613237 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.613202 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:14 crc kubenswrapper[4756]: E0203 09:11:14.613357 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:14 crc kubenswrapper[4756]: E0203 09:11:14.613593 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:14 crc kubenswrapper[4756]: E0203 09:11:14.613633 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:14 crc kubenswrapper[4756]: E0203 09:11:14.613707 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.619779 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.619844 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.619863 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.619889 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.619907 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.721755 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.721817 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.721836 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.721859 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.721876 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.823993 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.824569 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.824639 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.824738 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.824801 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.928117 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.928173 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.928186 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.928209 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:14 crc kubenswrapper[4756]: I0203 09:11:14.928227 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:14Z","lastTransitionTime":"2026-02-03T09:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.030791 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.030892 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.030904 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.030928 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.030940 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.133023 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.133069 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.133085 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.133102 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.133114 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.235874 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.235929 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.235938 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.235952 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.235961 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.337712 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.337753 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.337765 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.337781 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.337791 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.440607 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.440661 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.440670 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.440684 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.440692 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.543110 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.543156 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.543171 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.543188 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.543199 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.599890 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:28:32.559318787 +0000 UTC Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.645909 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.645952 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.645962 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.645976 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.645984 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.699342 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.699388 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.699396 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.699411 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.699419 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: E0203 09:11:15.719185 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:15Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.722691 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.722723 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.722735 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.722749 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.722758 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: E0203 09:11:15.734709 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:15Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.738613 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.738658 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.738670 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.738685 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.738697 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: E0203 09:11:15.755282 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:15Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.760475 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.760531 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.760550 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.760577 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.760598 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: E0203 09:11:15.783975 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:15Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.788047 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.788080 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.788090 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.788105 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.788117 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: E0203 09:11:15.803276 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:15Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:15 crc kubenswrapper[4756]: E0203 09:11:15.803608 4756 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.804759 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.804864 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.804877 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.804890 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.804898 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.907961 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.908032 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.908046 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.908075 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:15 crc kubenswrapper[4756]: I0203 09:11:15.908091 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:15Z","lastTransitionTime":"2026-02-03T09:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.009842 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.010070 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.010174 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.010262 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.010339 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.111925 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.111952 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.111960 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.111972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.111980 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.214048 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.214085 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.214093 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.214108 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.214117 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.316588 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.316622 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.316632 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.316648 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.316660 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.418642 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.418684 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.418702 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.418717 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.418726 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.520813 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.520860 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.520875 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.520895 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.520916 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.600275 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 16:37:01.169201581 +0000 UTC Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.613669 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.613694 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.613773 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.614000 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:16 crc kubenswrapper[4756]: E0203 09:11:16.613989 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:16 crc kubenswrapper[4756]: E0203 09:11:16.614076 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:16 crc kubenswrapper[4756]: E0203 09:11:16.614206 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:16 crc kubenswrapper[4756]: E0203 09:11:16.614267 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.623528 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.623561 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.623580 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.623596 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.623609 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.726077 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.726128 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.726137 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.726152 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.726164 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.829139 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.829184 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.829201 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.829221 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.829237 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.932186 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.932229 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.932238 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.932252 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.932261 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:16Z","lastTransitionTime":"2026-02-03T09:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.985909 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/0.log" Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.986006 4756 generic.go:334] "Generic (PLEG): container finished" podID="89e5d60f-f690-486f-b078-5ae9f98e1f3a" containerID="bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577" exitCode=1 Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.986037 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6srkq" event={"ID":"89e5d60f-f690-486f-b078-5ae9f98e1f3a","Type":"ContainerDied","Data":"bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577"} Feb 03 09:11:16 crc kubenswrapper[4756]: I0203 09:11:16.986529 4756 scope.go:117] "RemoveContainer" containerID="bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.006044 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.019123 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.042612 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.042817 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.042852 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.042863 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.042877 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.042886 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.054508 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.066950 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"2026-02-03T09:10:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c\\\\n2026-02-03T09:10:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c to /host/opt/cni/bin/\\\\n2026-02-03T09:10:31Z [verbose] multus-daemon started\\\\n2026-02-03T09:10:31Z [verbose] Readiness Indicator file check\\\\n2026-02-03T09:11:16Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.088111 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.102777 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.115653 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.132768 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.145932 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.145982 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.145993 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.146009 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.146024 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.151711 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.167100 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.181508 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.194198 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.204563 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.214338 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.226388 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.236731 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:17Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.248092 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.248116 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.248125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.248138 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.248146 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.351673 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.351720 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.351731 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.351748 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.351758 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.456127 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.456181 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.456193 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.456210 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.456222 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.560025 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.560086 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.560098 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.560111 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.560122 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.600541 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 00:39:42.4938103 +0000 UTC Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.664036 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.664125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.664140 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.664163 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.664176 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.767079 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.767115 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.767125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.767138 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.767153 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.869981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.870022 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.870033 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.870049 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.870060 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.972341 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.972368 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.972377 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.972390 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.972398 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:17Z","lastTransitionTime":"2026-02-03T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.991400 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/0.log" Feb 03 09:11:17 crc kubenswrapper[4756]: I0203 09:11:17.991470 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6srkq" event={"ID":"89e5d60f-f690-486f-b078-5ae9f98e1f3a","Type":"ContainerStarted","Data":"d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.008797 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.021704 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.031583 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.043057 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.054671 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.066589 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.074807 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.074842 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.074852 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.074870 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.074879 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.079334 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.090343 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.102663 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.116095 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.127390 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.139401 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.159519 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.171477 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.177578 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.177616 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.177628 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.177643 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.177654 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.184833 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.200201 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.214896 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"2026-02-03T09:10:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c\\\\n2026-02-03T09:10:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c to /host/opt/cni/bin/\\\\n2026-02-03T09:10:31Z [verbose] multus-daemon started\\\\n2026-02-03T09:10:31Z [verbose] Readiness Indicator file check\\\\n2026-02-03T09:11:16Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:18Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.279704 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.279769 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.279781 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.279799 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.279810 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.382848 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.382903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.382920 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.382945 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.382962 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.485483 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.485527 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.485540 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.485553 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.485564 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.588283 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.588320 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.588331 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.588347 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.588358 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.601501 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 23:09:47.791322402 +0000 UTC Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.613959 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.613977 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.614010 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:18 crc kubenswrapper[4756]: E0203 09:11:18.614067 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.614117 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:18 crc kubenswrapper[4756]: E0203 09:11:18.614137 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:18 crc kubenswrapper[4756]: E0203 09:11:18.614198 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:18 crc kubenswrapper[4756]: E0203 09:11:18.614407 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.691582 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.691620 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.691631 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.691648 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.691659 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.793911 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.793953 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.793963 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.793979 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.793989 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.896183 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.896272 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.896285 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.896304 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.896314 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.998386 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.998429 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.998458 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.998474 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:18 crc kubenswrapper[4756]: I0203 09:11:18.998483 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:18Z","lastTransitionTime":"2026-02-03T09:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.101168 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.101217 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.101227 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.101244 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.101256 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.204028 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.204061 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.204070 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.204085 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.204096 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.306271 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.306305 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.306315 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.306328 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.306336 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.408601 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.408656 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.408670 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.408689 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.408710 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.510678 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.510977 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.511087 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.511192 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.511270 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.602559 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 23:49:50.626446792 +0000 UTC Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.615602 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.615748 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.615830 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.615918 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.615999 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.625124 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.637124 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.648757 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"2026-02-03T09:10:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c\\\\n2026-02-03T09:10:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c to /host/opt/cni/bin/\\\\n2026-02-03T09:10:31Z [verbose] multus-daemon started\\\\n2026-02-03T09:10:31Z [verbose] Readiness Indicator file check\\\\n2026-02-03T09:11:16Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.669423 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.680862 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.695164 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.706424 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.717386 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.717681 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.717864 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.717958 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.718039 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.718101 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.730101 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.741755 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.751483 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.763947 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.779386 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.794123 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.806577 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.818240 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.820898 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.820974 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.820989 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.821010 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.821022 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.827163 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:19Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.922959 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.923002 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.923013 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.923028 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:19 crc kubenswrapper[4756]: I0203 09:11:19.923037 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:19Z","lastTransitionTime":"2026-02-03T09:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.025208 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.025244 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.025254 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.025304 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.025317 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.127821 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.127866 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.127879 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.127895 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.127907 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.230562 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.230609 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.230617 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.230633 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.230642 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.332913 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.332946 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.332956 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.332971 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.332983 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.434725 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.434992 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.435059 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.435131 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.435191 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.538707 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.538748 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.538758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.538775 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.538787 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.603149 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 15:51:50.240211503 +0000 UTC Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.613115 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.613222 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.613320 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:20 crc kubenswrapper[4756]: E0203 09:11:20.613319 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.613500 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:20 crc kubenswrapper[4756]: E0203 09:11:20.613499 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:20 crc kubenswrapper[4756]: E0203 09:11:20.613623 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:20 crc kubenswrapper[4756]: E0203 09:11:20.613725 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.640915 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.640954 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.640964 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.640978 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.640988 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.742568 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.742605 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.742616 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.742631 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.742641 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.844568 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.844599 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.844609 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.844624 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.844635 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.947066 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.947103 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.947111 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.947125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:20 crc kubenswrapper[4756]: I0203 09:11:20.947133 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:20Z","lastTransitionTime":"2026-02-03T09:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.049047 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.049083 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.049091 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.049103 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.049113 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.151068 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.151099 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.151109 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.151121 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.151130 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.253372 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.253406 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.253415 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.253433 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.253459 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.355324 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.355365 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.355376 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.355392 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.355404 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.469719 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.469753 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.469764 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.469778 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.469790 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.571735 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.571877 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.571903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.571924 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.571945 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.604096 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 20:14:57.2004928 +0000 UTC Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.673532 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.673568 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.673579 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.673623 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.673639 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.775957 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.776005 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.776019 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.776035 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.776045 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.878070 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.878118 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.878132 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.878153 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.878168 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.980771 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.980813 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.980827 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.980844 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:21 crc kubenswrapper[4756]: I0203 09:11:21.980855 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:21Z","lastTransitionTime":"2026-02-03T09:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.083710 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.083748 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.083758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.083774 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.083787 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.186831 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.186887 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.186900 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.186924 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.186943 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.289385 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.289439 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.289469 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.289487 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.289497 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.392568 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.392634 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.392644 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.392658 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.392667 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.494961 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.495038 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.495051 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.495072 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.495092 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.598058 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.598113 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.598124 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.598142 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.598155 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.604243 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 15:36:14.590171493 +0000 UTC Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.613681 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.613693 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.613703 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.613690 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:22 crc kubenswrapper[4756]: E0203 09:11:22.613770 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:22 crc kubenswrapper[4756]: E0203 09:11:22.613875 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:22 crc kubenswrapper[4756]: E0203 09:11:22.613959 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:22 crc kubenswrapper[4756]: E0203 09:11:22.614015 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.614597 4756 scope.go:117] "RemoveContainer" containerID="9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.699913 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.699973 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.699984 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.700000 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.700012 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.801835 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.801876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.801889 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.801912 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.801928 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.904095 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.904135 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.904148 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.904178 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:22 crc kubenswrapper[4756]: I0203 09:11:22.904188 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:22Z","lastTransitionTime":"2026-02-03T09:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.006827 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.006855 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.006864 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.006877 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.006887 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.010001 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/2.log" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.013161 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.014501 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.031671 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.054492 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.072861 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.096514 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.109700 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.109730 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.109738 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.109751 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.109759 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.109737 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"2026-02-03T09:10:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c\\\\n2026-02-03T09:10:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c to /host/opt/cni/bin/\\\\n2026-02-03T09:10:31Z [verbose] multus-daemon started\\\\n2026-02-03T09:10:31Z [verbose] Readiness Indicator file check\\\\n2026-02-03T09:11:16Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.142699 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.158576 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.170351 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.181385 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.195274 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.206461 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.212063 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.212090 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.212097 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.212110 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.212118 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.240157 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.255964 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.280296 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.294732 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.309531 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.313822 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.313865 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.313876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.313895 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.313908 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.321543 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.416267 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.416310 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.416320 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.416334 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.416343 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.519001 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.519045 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.519054 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.519068 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.519077 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.604504 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 15:24:45.946156234 +0000 UTC Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.620547 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.620576 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.620585 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.620597 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.620604 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.722586 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.722635 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.722647 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.722675 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.722692 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.825527 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.825568 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.825578 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.825594 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.825604 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.928141 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.928216 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.928227 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.928268 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:23 crc kubenswrapper[4756]: I0203 09:11:23.928281 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:23Z","lastTransitionTime":"2026-02-03T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.018263 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/3.log" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.019089 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/2.log" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.021879 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" exitCode=1 Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.021915 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.021947 4756 scope.go:117] "RemoveContainer" containerID="9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.022898 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:11:24 crc kubenswrapper[4756]: E0203 09:11:24.023112 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.031274 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.031310 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.031318 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.031330 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.031340 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.036882 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.049050 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.061371 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.074824 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.091916 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bde3bd9ba24a7fe0b15161aa0d89199ec906aa81765bda9607cf96c8b674138\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:10:56Z\\\",\\\"message\\\":\\\" 6449 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0203 09:10:56.462264 6449 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0203 09:10:56.462285 6449 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0203 09:10:56.462290 6449 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0203 09:10:56.462310 6449 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0203 09:10:56.462312 6449 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0203 09:10:56.462319 6449 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0203 09:10:56.462361 6449 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0203 09:10:56.462373 6449 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0203 09:10:56.462393 6449 factory.go:656] Stopping watch factory\\\\nI0203 09:10:56.462407 6449 handler.go:208] Removed *v1.Node event handler 7\\\\nI0203 09:10:56.462415 6449 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0203 09:10:56.462432 6449 handler.go:208] Removed *v1.Node event handler 2\\\\nI0203 09:10:56.462434 6449 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0203 09:10:56.462593 6449 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:23Z\\\",\\\"message\\\":\\\"t network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0203 09:11:23.567730 6847 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0203 09:11:23.567735 6847 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}\\\\nI0203 09:11:23.567743 6847 services_controller.go:360] Finished syncing service multus-admission-controller on namespace openshift-multus for network=default : 2.413721ms\\\\nI0203 09:11:23.567748 6847 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.103527 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.115262 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.128717 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.141456 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.141492 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.141500 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.141512 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.141523 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.142418 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"2026-02-03T09:10:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c\\\\n2026-02-03T09:10:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c to /host/opt/cni/bin/\\\\n2026-02-03T09:10:31Z [verbose] multus-daemon started\\\\n2026-02-03T09:10:31Z [verbose] Readiness Indicator file check\\\\n2026-02-03T09:11:16Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.154373 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.167382 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.176092 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.186280 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.200653 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.218717 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.233291 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.244288 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.244348 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.244363 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.244432 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.244481 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.245698 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:24Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.346664 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.346690 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.346697 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.346709 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.346723 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.448421 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.448464 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.448472 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.448485 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.448494 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.550514 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.550554 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.550564 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.550579 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.550588 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.605491 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 19:49:21.458616245 +0000 UTC Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.613828 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.613851 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.613893 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:24 crc kubenswrapper[4756]: E0203 09:11:24.613957 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.613979 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:24 crc kubenswrapper[4756]: E0203 09:11:24.614094 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:24 crc kubenswrapper[4756]: E0203 09:11:24.614171 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:24 crc kubenswrapper[4756]: E0203 09:11:24.614231 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.652664 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.652691 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.652698 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.652709 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.652717 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.755683 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.755721 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.755757 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.755771 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.755784 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.857549 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.857577 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.857585 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.857597 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.857605 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.960400 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.960532 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.960558 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.960585 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:24 crc kubenswrapper[4756]: I0203 09:11:24.960604 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:24Z","lastTransitionTime":"2026-02-03T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.027496 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/3.log" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.032836 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:11:25 crc kubenswrapper[4756]: E0203 09:11:25.034013 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.050525 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.063631 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.063667 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.063675 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.063688 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.063699 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.064595 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.078658 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.095406 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.106855 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.117961 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.126831 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.135037 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.146567 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.156934 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.168226 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.168266 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.168274 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.168288 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.168298 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.170733 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.187692 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.201004 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.213629 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.225940 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"2026-02-03T09:10:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c\\\\n2026-02-03T09:10:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c to /host/opt/cni/bin/\\\\n2026-02-03T09:10:31Z [verbose] multus-daemon started\\\\n2026-02-03T09:10:31Z [verbose] Readiness Indicator file check\\\\n2026-02-03T09:11:16Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.244174 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:23Z\\\",\\\"message\\\":\\\"t network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0203 09:11:23.567730 6847 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0203 09:11:23.567735 6847 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}\\\\nI0203 09:11:23.567743 6847 services_controller.go:360] Finished syncing service multus-admission-controller on namespace openshift-multus for network=default : 2.413721ms\\\\nI0203 09:11:23.567748 6847 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:11:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.255114 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:25Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.271217 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.271268 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.271283 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.271301 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.271312 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.374020 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.374081 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.374104 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.374124 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.374139 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.475706 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.475754 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.475768 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.475783 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.475793 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.577780 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.577831 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.577843 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.577860 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.577872 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.606429 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 18:19:07.965001598 +0000 UTC Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.680866 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.680915 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.680926 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.680944 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.680956 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.784118 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.784165 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.784176 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.784194 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.784208 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.886840 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.886931 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.886948 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.886970 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.886988 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.990944 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.991005 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.991021 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.991043 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:25 crc kubenswrapper[4756]: I0203 09:11:25.991069 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:25Z","lastTransitionTime":"2026-02-03T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.000159 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.000232 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.000255 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.000274 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.000288 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.022328 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:26Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.026488 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.026537 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.026550 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.026566 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.026577 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.041416 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:26Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.045495 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.045540 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.045555 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.045575 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.045592 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.064096 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:26Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.069115 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.069165 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.069175 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.069193 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.069205 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.084041 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:26Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.088151 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.088194 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.088209 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.088233 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.088247 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.107124 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:26Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.107295 4756 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.109405 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.109481 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.109499 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.109520 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.109533 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.212490 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.212552 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.212569 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.212592 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.212609 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.315844 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.315897 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.315909 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.315927 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.315939 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.419094 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.419168 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.419181 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.419203 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.419222 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.522852 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.522907 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.522932 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.522963 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.522983 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.607624 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 18:05:03.249973816 +0000 UTC Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.613138 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.613184 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.613211 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.613154 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.613412 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.613628 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.613804 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:26 crc kubenswrapper[4756]: E0203 09:11:26.613997 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.625945 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.625997 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.626015 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.626036 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.626054 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.729048 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.729127 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.729140 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.729166 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.729182 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.831533 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.831619 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.831642 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.831670 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.831692 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.933765 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.933829 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.933846 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.933868 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:26 crc kubenswrapper[4756]: I0203 09:11:26.933885 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:26Z","lastTransitionTime":"2026-02-03T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.036416 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.036483 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.036497 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.036512 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.036524 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.139711 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.139759 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.139774 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.139791 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.139802 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.243084 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.243124 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.243135 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.243152 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.243165 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.346071 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.346128 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.346147 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.346169 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.346185 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.449117 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.449168 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.449177 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.449192 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.449203 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.551266 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.551347 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.551359 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.551391 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.551402 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.608776 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 20:32:05.491713064 +0000 UTC Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.654044 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.654094 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.654113 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.654148 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.654164 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.760401 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.760476 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.760493 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.760511 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.760522 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.864432 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.864517 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.864528 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.864562 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.864577 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.967928 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.967990 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.968005 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.968031 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:27 crc kubenswrapper[4756]: I0203 09:11:27.968049 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:27Z","lastTransitionTime":"2026-02-03T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.070799 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.070873 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.070890 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.070916 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.070932 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.174019 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.174081 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.174095 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.174118 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.174137 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.277698 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.277785 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.277808 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.277842 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.277868 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.381799 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.381876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.381893 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.381916 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.381934 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.485405 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.485477 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.485492 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.485514 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.485527 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.589983 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.590026 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.590037 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.590056 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.590068 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.609964 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 10:01:34.592217819 +0000 UTC Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.613289 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.613314 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.613362 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.613348 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:28 crc kubenswrapper[4756]: E0203 09:11:28.613463 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:28 crc kubenswrapper[4756]: E0203 09:11:28.613532 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:28 crc kubenswrapper[4756]: E0203 09:11:28.613595 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:28 crc kubenswrapper[4756]: E0203 09:11:28.613714 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.693206 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.693287 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.693329 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.693350 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.693365 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.795761 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.795800 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.795811 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.795827 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.795838 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.898488 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.898524 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.898536 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.898552 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:28 crc kubenswrapper[4756]: I0203 09:11:28.898563 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:28Z","lastTransitionTime":"2026-02-03T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.001618 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.001690 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.001701 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.001718 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.001729 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.104903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.104969 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.104986 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.105008 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.105026 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.208259 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.208350 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.208365 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.208388 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.208403 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.311704 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.311760 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.311778 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.311801 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.311818 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.414753 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.414825 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.414843 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.414866 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.414884 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.517656 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.517721 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.517739 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.517764 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.517782 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.610680 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 03:38:03.795921052 +0000 UTC Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.620082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.620159 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.620177 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.620207 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.620226 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.630225 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.648317 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.661484 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.676534 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.693817 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.715794 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"2026-02-03T09:10:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c\\\\n2026-02-03T09:10:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c to /host/opt/cni/bin/\\\\n2026-02-03T09:10:31Z [verbose] multus-daemon started\\\\n2026-02-03T09:10:31Z [verbose] Readiness Indicator file check\\\\n2026-02-03T09:11:16Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.722370 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.722410 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.722419 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.722433 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.722460 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.742590 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:23Z\\\",\\\"message\\\":\\\"t network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0203 09:11:23.567730 6847 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0203 09:11:23.567735 6847 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}\\\\nI0203 09:11:23.567743 6847 services_controller.go:360] Finished syncing service multus-admission-controller on namespace openshift-multus for network=default : 2.413721ms\\\\nI0203 09:11:23.567748 6847 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:11:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.754606 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.765264 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.776684 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.790237 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.806517 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.825578 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.825911 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.825949 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.825958 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.825972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.825981 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.842976 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.860414 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.873964 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.887262 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:29Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.928747 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.928783 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.928793 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.928808 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:29 crc kubenswrapper[4756]: I0203 09:11:29.928821 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:29Z","lastTransitionTime":"2026-02-03T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.031394 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.031473 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.031486 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.031504 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.031519 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.133949 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.133989 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.134005 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.134024 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.134038 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.237367 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.237436 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.237473 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.237492 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.237516 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.340923 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.340965 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.341002 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.341019 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.341030 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.443335 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.443374 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.443384 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.443431 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.443505 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.546985 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.547068 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.547078 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.547092 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.547125 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.611868 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 22:48:30.383900482 +0000 UTC Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.613207 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.613296 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.613298 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.613208 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:30 crc kubenswrapper[4756]: E0203 09:11:30.613428 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:30 crc kubenswrapper[4756]: E0203 09:11:30.613599 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:30 crc kubenswrapper[4756]: E0203 09:11:30.613731 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:30 crc kubenswrapper[4756]: E0203 09:11:30.613826 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.626880 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.649784 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.649859 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.649882 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.649913 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.649936 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.752694 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.752729 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.752738 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.752750 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.752759 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.855487 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.855535 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.855546 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.855564 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.855575 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.957730 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.957776 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.957794 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.957811 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:30 crc kubenswrapper[4756]: I0203 09:11:30.957820 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:30Z","lastTransitionTime":"2026-02-03T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.060809 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.060872 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.060895 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.060926 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.060947 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.164125 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.164173 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.164185 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.164202 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.164214 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.266658 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.266823 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.266858 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.266887 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.266911 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.370693 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.370772 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.370796 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.370831 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.370854 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.474277 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.474323 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.474337 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.474356 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.474369 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.577185 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.577245 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.577265 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.577293 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.577316 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.612819 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 18:49:48.582040077 +0000 UTC Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.680018 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.680069 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.680082 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.680104 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.680118 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.783270 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.783351 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.783372 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.783396 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.783495 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.885431 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.885500 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.885518 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.885545 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.885558 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.988645 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.988739 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.988776 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.988806 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:31 crc kubenswrapper[4756]: I0203 09:11:31.988828 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:31Z","lastTransitionTime":"2026-02-03T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.091248 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.091295 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.091343 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.091362 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.091376 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.193856 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.193889 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.193899 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.193914 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.193924 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.289121 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.289362 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.289335304 +0000 UTC m=+147.439802679 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.297711 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.297756 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.297767 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.297786 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.297797 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.390305 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.390382 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.390422 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390503 4756 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.390533 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390546 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390594 4756 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390619 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390642 4756 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390662 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390674 4756 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390683 4756 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.390587 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.390569981 +0000 UTC m=+147.541037356 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.391094 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.391081688 +0000 UTC m=+147.541549063 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.391121 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.391113429 +0000 UTC m=+147.541580804 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.391131 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.39112645 +0000 UTC m=+147.541593825 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.399402 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.399432 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.399468 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.399485 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.399498 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.502048 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.502106 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.502123 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.502144 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.502163 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.605088 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.605165 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.605189 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.605220 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.605242 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.613306 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 06:44:54.333722735 +0000 UTC Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.613432 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.613504 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.613520 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.613602 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.613653 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.613787 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.613849 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:32 crc kubenswrapper[4756]: E0203 09:11:32.613967 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.707833 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.707860 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.707868 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.707881 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.707890 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.810727 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.810776 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.810787 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.810802 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.810815 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.913391 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.913494 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.913516 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.913549 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:32 crc kubenswrapper[4756]: I0203 09:11:32.913568 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:32Z","lastTransitionTime":"2026-02-03T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.016512 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.016644 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.016671 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.016703 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.016726 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.119730 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.119840 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.119911 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.119945 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.119965 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.223946 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.224021 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.224044 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.224074 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.224096 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.327109 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.327173 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.327189 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.327208 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.327222 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.430165 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.430664 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.430682 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.430710 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.430740 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.534943 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.534998 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.535014 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.535037 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.535053 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.614076 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 22:11:26.464691729 +0000 UTC Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.637484 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.637544 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.637562 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.637583 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.637599 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.740877 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.740983 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.741021 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.741056 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.741079 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.843846 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.843906 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.843925 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.843949 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.843969 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.947032 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.947092 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.947110 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.947134 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:33 crc kubenswrapper[4756]: I0203 09:11:33.947151 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:33Z","lastTransitionTime":"2026-02-03T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.049350 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.049530 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.049572 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.049604 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.049627 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.153863 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.154001 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.154029 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.154061 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.154082 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.257590 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.257658 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.257680 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.257706 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.257727 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.365522 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.365573 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.365581 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.365595 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.365610 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.468285 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.468320 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.468329 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.468341 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.468349 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.570812 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.571245 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.571273 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.571293 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.571310 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.613557 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.613610 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.613663 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:34 crc kubenswrapper[4756]: E0203 09:11:34.613705 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.613560 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:34 crc kubenswrapper[4756]: E0203 09:11:34.613828 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:34 crc kubenswrapper[4756]: E0203 09:11:34.613856 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:34 crc kubenswrapper[4756]: E0203 09:11:34.613894 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.614226 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 06:42:59.688111544 +0000 UTC Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.673880 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.674412 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.674576 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.674669 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.674738 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.776728 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.776797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.776806 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.776821 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.776830 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.878779 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.878836 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.878846 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.878862 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.878871 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.981312 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.981356 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.981365 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.981384 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:34 crc kubenswrapper[4756]: I0203 09:11:34.981394 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:34Z","lastTransitionTime":"2026-02-03T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.083667 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.083733 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.083750 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.083776 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.083794 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.186068 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.186134 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.186156 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.186185 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.186206 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.289527 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.289577 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.289595 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.289618 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.289634 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.392325 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.392394 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.392416 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.392473 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.392498 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.494517 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.494547 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.494555 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.494567 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.494576 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.597632 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.597688 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.597700 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.597717 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.597729 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.615374 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 13:59:58.270432144 +0000 UTC Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.700256 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.700291 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.700395 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.700411 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.700467 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.803487 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.803542 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.803557 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.803579 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.803594 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.905848 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.905902 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.905920 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.905940 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:35 crc kubenswrapper[4756]: I0203 09:11:35.905955 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:35Z","lastTransitionTime":"2026-02-03T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.008342 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.008389 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.008400 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.008415 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.008426 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.111110 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.111186 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.111209 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.111238 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.111261 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.213797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.213862 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.213886 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.213915 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.213937 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.289217 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.289290 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.289313 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.289342 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.289367 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.305830 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.310233 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.310273 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.310286 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.310302 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.310314 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.328855 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.333519 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.333575 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.333587 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.333604 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.333617 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.352135 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.356477 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.356648 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.356750 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.356876 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.357030 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.371538 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.377847 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.377906 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.377925 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.377944 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.377959 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.389543 4756 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5e8183e3-1710-4801-877d-2bd18fb91137\\\",\\\"systemUUID\\\":\\\"8417792e-76c4-4111-84f0-a989146d0caa\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:36Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.389693 4756 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.391185 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.391218 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.391230 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.391245 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.391257 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.494296 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.494427 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.494476 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.494495 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.494508 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.597260 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.597311 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.597327 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.597348 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.597364 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.613685 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.613729 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.613753 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.613685 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.613838 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.613904 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.613989 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:36 crc kubenswrapper[4756]: E0203 09:11:36.614099 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.616237 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 18:16:27.268727247 +0000 UTC Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.699160 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.699194 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.699206 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.699226 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.699237 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.800941 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.800990 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.801005 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.801025 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.801040 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.904237 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.904286 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.904296 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.904311 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:36 crc kubenswrapper[4756]: I0203 09:11:36.904324 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:36Z","lastTransitionTime":"2026-02-03T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.007740 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.007807 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.007823 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.007848 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.007861 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.111338 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.111397 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.111426 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.111477 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.111496 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.214720 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.214797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.214822 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.214852 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.214876 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.317931 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.318319 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.318506 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.318663 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.318804 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.421143 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.421192 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.421206 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.421228 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.421243 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.523731 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.524151 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.524223 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.524287 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.524381 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.616373 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 10:01:31.806594534 +0000 UTC Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.626557 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.626615 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.626632 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.626655 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.626671 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.729292 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.729364 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.729387 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.729435 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.729552 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.832867 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.833049 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.833091 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.833129 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.833165 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.936499 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.936591 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.936610 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.936634 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:37 crc kubenswrapper[4756]: I0203 09:11:37.936651 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:37Z","lastTransitionTime":"2026-02-03T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.039903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.039956 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.039965 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.039980 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.039990 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.141913 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.141941 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.141950 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.141964 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.141973 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.244857 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.244899 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.244909 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.244925 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.244936 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.347200 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.347280 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.347307 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.347340 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.347364 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.450086 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.450126 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.450136 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.450152 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.450163 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.552037 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.552069 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.552077 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.552090 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.552098 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.613641 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.613693 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.613737 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.613751 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:38 crc kubenswrapper[4756]: E0203 09:11:38.614528 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:38 crc kubenswrapper[4756]: E0203 09:11:38.614639 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:38 crc kubenswrapper[4756]: E0203 09:11:38.614827 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:38 crc kubenswrapper[4756]: E0203 09:11:38.614923 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.616740 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 23:11:12.395473307 +0000 UTC Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.655066 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.655127 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.655146 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.655174 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.655191 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.758221 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.758283 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.758300 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.758325 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.758347 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.861966 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.862043 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.862068 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.862101 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.862125 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.968784 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.969155 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.969491 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.970247 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:38 crc kubenswrapper[4756]: I0203 09:11:38.970427 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:38Z","lastTransitionTime":"2026-02-03T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.073571 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.073898 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.074044 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.074176 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.074305 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.177040 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.177562 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.177663 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.177768 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.177854 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.281558 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.281950 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.282059 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.282152 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.282237 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.385092 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.385435 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.385647 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.385809 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.385956 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.489465 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.489946 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.490152 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.490302 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.490545 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.593887 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.593954 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.593975 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.594002 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.594020 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.614129 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:11:39 crc kubenswrapper[4756]: E0203 09:11:39.614280 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.616932 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 05:52:03.690613293 +0000 UTC Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.631419 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.650996 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6a6a991-ccf4-409f-bda5-cabc5788ea88\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5282fb5b329e8d2eb1c44fcd69c48e08a69e9d44e05458dbcb3216a9363bd126\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://899e8118651c662d6c9f7b56a899f70831d72814bece0fea028c49031b5aa69c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c58dbdcd6bdc7608f1e5210199d0f55257e7dada2c1b633ffb0c4f7f093cf33\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82b0a282b357a49ef75d3cf8b5072e956faa302f9fac8abe939f1cdb9ea2f6ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26394dcb7f7410a8dce645fb0c01f1fd4805eaa0950f2b8846e82fb9cb6f2b73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44fbfd8c2e261f3bb3d821c3aa1b202a2e521dbec24fd6e5933e04c4771940dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7998c8be25bcbb816baa4e0b316319d03f43ad65a38448911ee48ee9030519bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6ld46\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qx7sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.668749 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e9cfd18-1b1c-4c27-a693-22868b869042\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53ac2aa435dc240038511fcb4082a363758612ab23769f9108dd4d5646782dcd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://06c68e5158056b8a13cf2aa0593dc8e0f1ccee551670e399450a87f8c8503f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p2vjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lc4vl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.687701 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3fa3c91-71ed-4e18-9828-2f1885a87768\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5e9bc9f4527ae726293244f401b7c184a2b35b20769d83cc1a5c110d8006166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7cad5e50551f6e8e05ce661748c332abf4b30ba640cffa6a0021eb864000ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0d8ca369a2fb78fbe1b40cc55ba300ed5c968fad222c81d9fa7ecf1485cabfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb0c1494af70fd94ab4f95dfbde287513c476fb071a9193b77b2b636d30cc57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.697271 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.697315 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.697326 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.697343 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.697355 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.704577 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66b88c73-0891-4403-86d6-8d7bfe1bc750\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb345ac4f3c06479f5dffe90991140eb445c45550b2958c95e38e23451f9c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d73bf3db31948c8a05591d65a087d4e068bd35d6728fc117e1dd32f6570dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80d73bf3db31948c8a05591d65a087d4e068bd35d6728fc117e1dd32f6570dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.718866 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.732271 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6srkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89e5d60f-f690-486f-b078-5ae9f98e1f3a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:16Z\\\",\\\"message\\\":\\\"2026-02-03T09:10:30+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c\\\\n2026-02-03T09:10:30+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b35d332e-dcf3-4305-b269-ece40f26bf5c to /host/opt/cni/bin/\\\\n2026-02-03T09:10:31Z [verbose] multus-daemon started\\\\n2026-02-03T09:10:31Z [verbose] Readiness Indicator file check\\\\n2026-02-03T09:11:16Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ckdq2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6srkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.752832 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-03T09:11:23Z\\\",\\\"message\\\":\\\"t network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0203 09:11:23.567730 6847 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI0203 09:11:23.567735 6847 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}\\\\nI0203 09:11:23.567743 6847 services_controller.go:360] Finished syncing service multus-admission-controller on namespace openshift-multus for network=default : 2.413721ms\\\\nI0203 09:11:23.567748 6847 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:11:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h849g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gfsqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.764322 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"748779a5-a5e9-4451-839c-805686b764c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afd5546b65cf0635d9ccd59406b473ff23a6d492aa00f01db885757d829e1dc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wflqj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-c9rn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.775430 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-cnrvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291b3e7-d0e4-4800-b6bb-9fe15140ab0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ae83170668aa9a0bfddada858a6003bc3edb8a8aaa7e082075cf214c53ae963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vz2xg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-cnrvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.786527 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"717b7410-dd64-44cb-ba9b-3436d82ebb95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-99zjn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k6pzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.800255 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.800319 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.800337 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.800360 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.800378 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.800376 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"04a454b2-1e04-4b51-a751-8f54cca1ae85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3624bd2d4b43ef9b0f186f781103bed8633a6f207bb334a75900f5631a53d1c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cd59c90c94ecd69a918799b47474157973eeb76a3c645294e5f320b921b5451\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68197ca3638b5e376766a95646af6950499fe14748bc7dad24e62d0b279896c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.819390 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c245f03f-c294-456f-915e-a86a4e666fc8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"message\\\":\\\"le observer\\\\nW0203 09:10:29.267604 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0203 09:10:29.267727 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0203 09:10:29.268707 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2554252349/tls.crt::/tmp/serving-cert-2554252349/tls.key\\\\\\\"\\\\nI0203 09:10:29.641058 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0203 09:10:29.645238 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0203 09:10:29.645349 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0203 09:10:29.645402 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0203 09:10:29.645433 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0203 09:10:29.657260 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0203 09:10:29.657289 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657294 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0203 09:10:29.657299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0203 09:10:29.657302 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0203 09:10:29.657305 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0203 09:10:29.657308 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0203 09:10:29.657573 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0203 09:10:29.662633 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-03T09:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-03T09:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:09Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.834673 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abf2b37777ab3f458dbfd12542c08937dd132c71251a492cf5aa1cfbb5462c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.853420 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eab9308b70a45c1e06b2895eb756decf7c9caffd6b0dd85378d9b62ebd8a55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.869441 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.881935 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f63f1d008565dea3d2827b5e39d2435628af382a3a190b1269b41beb062f7863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5305bbe549af9a17fa8e7a08df1960c98e0d75c5e9dd4223453fcf0c2a622d1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.890731 4756 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-626ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34e2d039-ed1a-4c41-a5ba-c94c7dbf79a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-03T09:10:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c41bd4fbd6eb7f1172e78c44b0f613ff2289dfa1ef96ee777c97df58efdea24c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-03T09:10:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2qq67\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-03T09:10:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-626ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-03T09:11:39Z is after 2025-08-24T17:21:41Z" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.902515 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.902614 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.902680 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.902736 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:39 crc kubenswrapper[4756]: I0203 09:11:39.902791 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:39Z","lastTransitionTime":"2026-02-03T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.004903 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.004943 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.004955 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.004972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.004984 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.107722 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.107792 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.107815 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.107843 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.107865 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.210670 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.210735 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.210750 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.210769 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.210783 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.313568 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.313836 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.313901 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.313972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.314041 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.416599 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.416654 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.416667 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.416686 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.416699 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.518673 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.518713 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.518721 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.518735 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.518744 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.613362 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.613362 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.613540 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.613558 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:40 crc kubenswrapper[4756]: E0203 09:11:40.613695 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:40 crc kubenswrapper[4756]: E0203 09:11:40.613814 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:40 crc kubenswrapper[4756]: E0203 09:11:40.613896 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:40 crc kubenswrapper[4756]: E0203 09:11:40.613935 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.617519 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 14:31:45.858656681 +0000 UTC Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.621195 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.621224 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.621232 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.621243 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.621252 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.723746 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.723780 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.723788 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.723800 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.723810 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.827016 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.827066 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.827081 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.827100 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.827113 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.928799 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.929401 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.929541 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.929635 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:40 crc kubenswrapper[4756]: I0203 09:11:40.929709 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:40Z","lastTransitionTime":"2026-02-03T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.031348 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.031798 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.031957 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.032100 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.032257 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.134568 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.134604 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.134613 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.134626 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.134634 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.237811 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.238274 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.238507 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.238806 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.238961 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.341693 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.341741 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.341754 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.341772 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.341783 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.444223 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.444310 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.444322 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.444338 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.444370 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.546930 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.546972 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.546981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.546997 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.547006 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.617609 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 23:10:33.963345872 +0000 UTC Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.649404 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.649436 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.649474 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.649489 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.649500 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.751710 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.751748 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.751758 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.751773 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.751784 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.854217 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.854260 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.854271 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.854287 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.854295 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.956784 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.956819 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.956832 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.956847 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:41 crc kubenswrapper[4756]: I0203 09:11:41.956858 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:41Z","lastTransitionTime":"2026-02-03T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.060212 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.060270 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.060282 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.060300 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.060312 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.163154 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.163206 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.163219 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.163236 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.163253 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.266153 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.266189 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.266197 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.266211 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.266220 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.368807 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.368852 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.368862 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.368879 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.368892 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.471099 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.471157 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.471171 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.471188 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.471203 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.573315 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.573430 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.573458 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.573471 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.573480 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.613625 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.613624 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.614046 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.614069 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:42 crc kubenswrapper[4756]: E0203 09:11:42.614221 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:42 crc kubenswrapper[4756]: E0203 09:11:42.614285 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:42 crc kubenswrapper[4756]: E0203 09:11:42.614373 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:42 crc kubenswrapper[4756]: E0203 09:11:42.614467 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.617790 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 17:47:23.198230421 +0000 UTC Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.676661 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.676702 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.676712 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.676725 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.676736 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.779534 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.779584 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.779595 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.779609 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.779623 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.881475 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.881513 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.881523 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.881538 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.881550 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.984245 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.984305 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.984321 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.984346 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:42 crc kubenswrapper[4756]: I0203 09:11:42.984363 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:42Z","lastTransitionTime":"2026-02-03T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.087324 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.087743 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.087918 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.088076 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.088225 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.191371 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.191419 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.191432 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.191466 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.191478 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.294090 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.294142 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.294151 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.294165 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.294175 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.397173 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.397233 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.397250 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.397274 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.397293 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.499842 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.499878 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.499889 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.499907 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.499917 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.602236 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.602293 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.602310 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.602337 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.602353 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.618498 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 01:06:45.0229392 +0000 UTC Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.705760 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.705813 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.705864 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.705888 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.705900 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.808893 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.809179 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.809283 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.809387 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.809595 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.911976 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.912254 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.912343 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.912424 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:43 crc kubenswrapper[4756]: I0203 09:11:43.912529 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:43Z","lastTransitionTime":"2026-02-03T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.014875 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.014916 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.014924 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.014937 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.014947 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.117228 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.117295 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.117305 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.117319 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.117329 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.219686 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.219724 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.219737 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.219752 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.219763 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.322267 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.322310 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.322322 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.322336 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.322348 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.424320 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.424368 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.424380 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.424399 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.424413 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.526929 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.526958 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.526967 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.526981 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.526995 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.614001 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.614016 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:44 crc kubenswrapper[4756]: E0203 09:11:44.614641 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.614057 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:44 crc kubenswrapper[4756]: E0203 09:11:44.614720 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.614050 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:44 crc kubenswrapper[4756]: E0203 09:11:44.614775 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:44 crc kubenswrapper[4756]: E0203 09:11:44.614517 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.619089 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 17:12:01.63369568 +0000 UTC Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.628872 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.628935 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.628952 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.628977 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.629007 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.731351 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.731401 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.731416 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.731433 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.731467 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.833961 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.834013 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.834025 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.834039 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.834048 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.935673 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.935716 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.935725 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.935738 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:44 crc kubenswrapper[4756]: I0203 09:11:44.935747 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:44Z","lastTransitionTime":"2026-02-03T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.038622 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.038673 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.038683 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.038699 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.038712 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.141488 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.141532 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.141545 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.141561 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.141572 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.244729 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.244839 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.244889 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.244917 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.244936 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.347492 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.347528 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.347539 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.347551 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.347560 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.450461 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.450503 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.450514 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.450529 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.450540 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.553598 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.553659 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.553672 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.553689 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.553700 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.619402 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 00:40:47.075047951 +0000 UTC Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.656399 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.656461 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.656471 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.656490 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.656501 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.758643 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.758752 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.758775 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.758797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.758814 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.861136 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.861193 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.861202 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.861216 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.861226 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.963652 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.963723 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.963791 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.964107 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:45 crc kubenswrapper[4756]: I0203 09:11:45.964158 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:45Z","lastTransitionTime":"2026-02-03T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.066797 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.066830 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.066840 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.066854 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.066864 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:46Z","lastTransitionTime":"2026-02-03T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.169172 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.169252 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.169262 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.169277 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.169287 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:46Z","lastTransitionTime":"2026-02-03T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.271586 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.271634 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.271646 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.271666 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.271678 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:46Z","lastTransitionTime":"2026-02-03T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.373857 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.373913 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.373930 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.373950 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.373965 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:46Z","lastTransitionTime":"2026-02-03T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.469547 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.469601 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.469621 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.469639 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.469652 4756 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-03T09:11:46Z","lastTransitionTime":"2026-02-03T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.512725 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs"] Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.513098 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.515098 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.516582 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.516757 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.517085 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.550393 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-626ms" podStartSLOduration=77.550369763 podStartE2EDuration="1m17.550369763s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.540261298 +0000 UTC m=+97.690728673" watchObservedRunningTime="2026-02-03 09:11:46.550369763 +0000 UTC m=+97.700837138" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.589326 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-qx7sz" podStartSLOduration=78.589306924 podStartE2EDuration="1m18.589306924s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.568905548 +0000 UTC m=+97.719372923" watchObservedRunningTime="2026-02-03 09:11:46.589306924 +0000 UTC m=+97.739774299" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.598752 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lc4vl" podStartSLOduration=77.598733947 podStartE2EDuration="1m17.598733947s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.598602232 +0000 UTC m=+97.749069607" watchObservedRunningTime="2026-02-03 09:11:46.598733947 +0000 UTC m=+97.749201322" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.610747 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.610728645 podStartE2EDuration="48.610728645s" podCreationTimestamp="2026-02-03 09:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.610653282 +0000 UTC m=+97.761120667" watchObservedRunningTime="2026-02-03 09:11:46.610728645 +0000 UTC m=+97.761196020" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.613078 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.613079 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.613108 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.613498 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:46 crc kubenswrapper[4756]: E0203 09:11:46.613584 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:46 crc kubenswrapper[4756]: E0203 09:11:46.613744 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:46 crc kubenswrapper[4756]: E0203 09:11:46.613912 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:46 crc kubenswrapper[4756]: E0203 09:11:46.614084 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.620620 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 05:38:29.416379639 +0000 UTC Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.620859 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.630666 4756 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.639120 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=16.639103856 podStartE2EDuration="16.639103856s" podCreationTimestamp="2026-02-03 09:11:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.621899615 +0000 UTC m=+97.772366980" watchObservedRunningTime="2026-02-03 09:11:46.639103856 +0000 UTC m=+97.789571231" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.653797 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6srkq" podStartSLOduration=78.653780262 podStartE2EDuration="1m18.653780262s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.653597676 +0000 UTC m=+97.804065051" watchObservedRunningTime="2026-02-03 09:11:46.653780262 +0000 UTC m=+97.804247647" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.664852 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cda7af56-1933-4cc9-917f-ff8347bb7f14-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.665168 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cda7af56-1933-4cc9-917f-ff8347bb7f14-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.665294 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cda7af56-1933-4cc9-917f-ff8347bb7f14-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.665487 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cda7af56-1933-4cc9-917f-ff8347bb7f14-service-ca\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.665583 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.665619 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cda7af56-1933-4cc9-917f-ff8347bb7f14-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: E0203 09:11:46.665767 4756 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:11:46 crc kubenswrapper[4756]: E0203 09:11:46.665851 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs podName:717b7410-dd64-44cb-ba9b-3436d82ebb95 nodeName:}" failed. No retries permitted until 2026-02-03 09:12:50.665832112 +0000 UTC m=+161.816299477 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs") pod "network-metrics-daemon-k6pzt" (UID: "717b7410-dd64-44cb-ba9b-3436d82ebb95") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.690565 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podStartSLOduration=78.690550842 podStartE2EDuration="1m18.690550842s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.689484587 +0000 UTC m=+97.839951982" watchObservedRunningTime="2026-02-03 09:11:46.690550842 +0000 UTC m=+97.841018217" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.710640 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-cnrvx" podStartSLOduration=78.710625037 podStartE2EDuration="1m18.710625037s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.701395511 +0000 UTC m=+97.851862886" watchObservedRunningTime="2026-02-03 09:11:46.710625037 +0000 UTC m=+97.861092402" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.741353 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.741332736 podStartE2EDuration="1m17.741332736s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.741053226 +0000 UTC m=+97.891520601" watchObservedRunningTime="2026-02-03 09:11:46.741332736 +0000 UTC m=+97.891800111" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.741878 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=76.741871804 podStartE2EDuration="1m16.741871804s" podCreationTimestamp="2026-02-03 09:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:46.724091404 +0000 UTC m=+97.874558789" watchObservedRunningTime="2026-02-03 09:11:46.741871804 +0000 UTC m=+97.892339179" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.766859 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cda7af56-1933-4cc9-917f-ff8347bb7f14-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.767106 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cda7af56-1933-4cc9-917f-ff8347bb7f14-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.767267 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cda7af56-1933-4cc9-917f-ff8347bb7f14-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.767407 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cda7af56-1933-4cc9-917f-ff8347bb7f14-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.767520 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cda7af56-1933-4cc9-917f-ff8347bb7f14-service-ca\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.767693 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cda7af56-1933-4cc9-917f-ff8347bb7f14-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.767831 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cda7af56-1933-4cc9-917f-ff8347bb7f14-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.768231 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cda7af56-1933-4cc9-917f-ff8347bb7f14-service-ca\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.773175 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cda7af56-1933-4cc9-917f-ff8347bb7f14-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.785150 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cda7af56-1933-4cc9-917f-ff8347bb7f14-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-v86vs\" (UID: \"cda7af56-1933-4cc9-917f-ff8347bb7f14\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:46 crc kubenswrapper[4756]: I0203 09:11:46.827673 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" Feb 03 09:11:47 crc kubenswrapper[4756]: I0203 09:11:47.100514 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" event={"ID":"cda7af56-1933-4cc9-917f-ff8347bb7f14","Type":"ContainerStarted","Data":"f5392288abb93614bf40dc034716b305172e154280a15d3eb061de23e8a8af66"} Feb 03 09:11:47 crc kubenswrapper[4756]: I0203 09:11:47.100567 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" event={"ID":"cda7af56-1933-4cc9-917f-ff8347bb7f14","Type":"ContainerStarted","Data":"d85037f761354016d070d158b59241119978db2f8c881801d6595e7636b028d8"} Feb 03 09:11:48 crc kubenswrapper[4756]: I0203 09:11:48.613278 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:48 crc kubenswrapper[4756]: E0203 09:11:48.613697 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:48 crc kubenswrapper[4756]: I0203 09:11:48.613286 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:48 crc kubenswrapper[4756]: E0203 09:11:48.613796 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:48 crc kubenswrapper[4756]: I0203 09:11:48.613291 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:48 crc kubenswrapper[4756]: I0203 09:11:48.613309 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:48 crc kubenswrapper[4756]: E0203 09:11:48.613870 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:48 crc kubenswrapper[4756]: E0203 09:11:48.613901 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:50 crc kubenswrapper[4756]: I0203 09:11:50.614117 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:50 crc kubenswrapper[4756]: E0203 09:11:50.614922 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:50 crc kubenswrapper[4756]: I0203 09:11:50.614169 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:50 crc kubenswrapper[4756]: E0203 09:11:50.615200 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:50 crc kubenswrapper[4756]: I0203 09:11:50.614150 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:50 crc kubenswrapper[4756]: E0203 09:11:50.615429 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:50 crc kubenswrapper[4756]: I0203 09:11:50.614186 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:50 crc kubenswrapper[4756]: E0203 09:11:50.615709 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:52 crc kubenswrapper[4756]: I0203 09:11:52.613752 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:52 crc kubenswrapper[4756]: E0203 09:11:52.614389 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:52 crc kubenswrapper[4756]: I0203 09:11:52.613880 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:52 crc kubenswrapper[4756]: E0203 09:11:52.614796 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:52 crc kubenswrapper[4756]: I0203 09:11:52.613843 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:52 crc kubenswrapper[4756]: E0203 09:11:52.615074 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:52 crc kubenswrapper[4756]: I0203 09:11:52.613902 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:52 crc kubenswrapper[4756]: E0203 09:11:52.615351 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:52 crc kubenswrapper[4756]: I0203 09:11:52.632572 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-v86vs" podStartSLOduration=84.632558387 podStartE2EDuration="1m24.632558387s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:47.114894145 +0000 UTC m=+98.265361550" watchObservedRunningTime="2026-02-03 09:11:52.632558387 +0000 UTC m=+103.783025762" Feb 03 09:11:52 crc kubenswrapper[4756]: I0203 09:11:52.632689 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 03 09:11:54 crc kubenswrapper[4756]: I0203 09:11:54.613281 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:54 crc kubenswrapper[4756]: I0203 09:11:54.614191 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:54 crc kubenswrapper[4756]: I0203 09:11:54.614200 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:54 crc kubenswrapper[4756]: E0203 09:11:54.614315 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:54 crc kubenswrapper[4756]: I0203 09:11:54.614405 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:54 crc kubenswrapper[4756]: I0203 09:11:54.614827 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:11:54 crc kubenswrapper[4756]: E0203 09:11:54.614553 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:54 crc kubenswrapper[4756]: E0203 09:11:54.614611 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:54 crc kubenswrapper[4756]: E0203 09:11:54.615074 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gfsqr_openshift-ovn-kubernetes(3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" Feb 03 09:11:54 crc kubenswrapper[4756]: E0203 09:11:54.615269 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:56 crc kubenswrapper[4756]: I0203 09:11:56.613654 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:56 crc kubenswrapper[4756]: I0203 09:11:56.613730 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:56 crc kubenswrapper[4756]: I0203 09:11:56.613662 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:56 crc kubenswrapper[4756]: E0203 09:11:56.613786 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:56 crc kubenswrapper[4756]: I0203 09:11:56.613681 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:56 crc kubenswrapper[4756]: E0203 09:11:56.613859 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:56 crc kubenswrapper[4756]: E0203 09:11:56.613955 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:56 crc kubenswrapper[4756]: E0203 09:11:56.614171 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:58 crc kubenswrapper[4756]: I0203 09:11:58.613259 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:11:58 crc kubenswrapper[4756]: E0203 09:11:58.613867 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:11:58 crc kubenswrapper[4756]: I0203 09:11:58.613345 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:11:58 crc kubenswrapper[4756]: E0203 09:11:58.614088 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:11:58 crc kubenswrapper[4756]: I0203 09:11:58.613293 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:11:58 crc kubenswrapper[4756]: E0203 09:11:58.614296 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:11:58 crc kubenswrapper[4756]: I0203 09:11:58.613358 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:11:58 crc kubenswrapper[4756]: E0203 09:11:58.614488 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:11:59 crc kubenswrapper[4756]: I0203 09:11:59.653775 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=7.653747945 podStartE2EDuration="7.653747945s" podCreationTimestamp="2026-02-03 09:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:11:59.652809134 +0000 UTC m=+110.803276549" watchObservedRunningTime="2026-02-03 09:11:59.653747945 +0000 UTC m=+110.804215380" Feb 03 09:12:00 crc kubenswrapper[4756]: I0203 09:12:00.613333 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:00 crc kubenswrapper[4756]: I0203 09:12:00.613533 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:00 crc kubenswrapper[4756]: I0203 09:12:00.613480 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:00 crc kubenswrapper[4756]: I0203 09:12:00.613408 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:00 crc kubenswrapper[4756]: E0203 09:12:00.613790 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:00 crc kubenswrapper[4756]: E0203 09:12:00.613896 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:00 crc kubenswrapper[4756]: E0203 09:12:00.614009 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:00 crc kubenswrapper[4756]: E0203 09:12:00.614120 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:02 crc kubenswrapper[4756]: I0203 09:12:02.613808 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:02 crc kubenswrapper[4756]: I0203 09:12:02.613890 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:02 crc kubenswrapper[4756]: E0203 09:12:02.613927 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:02 crc kubenswrapper[4756]: E0203 09:12:02.614034 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:02 crc kubenswrapper[4756]: I0203 09:12:02.613825 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:02 crc kubenswrapper[4756]: I0203 09:12:02.613802 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:02 crc kubenswrapper[4756]: E0203 09:12:02.614193 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:02 crc kubenswrapper[4756]: E0203 09:12:02.614132 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:03 crc kubenswrapper[4756]: I0203 09:12:03.153270 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/1.log" Feb 03 09:12:03 crc kubenswrapper[4756]: I0203 09:12:03.153856 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/0.log" Feb 03 09:12:03 crc kubenswrapper[4756]: I0203 09:12:03.153898 4756 generic.go:334] "Generic (PLEG): container finished" podID="89e5d60f-f690-486f-b078-5ae9f98e1f3a" containerID="d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b" exitCode=1 Feb 03 09:12:03 crc kubenswrapper[4756]: I0203 09:12:03.153927 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6srkq" event={"ID":"89e5d60f-f690-486f-b078-5ae9f98e1f3a","Type":"ContainerDied","Data":"d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b"} Feb 03 09:12:03 crc kubenswrapper[4756]: I0203 09:12:03.153960 4756 scope.go:117] "RemoveContainer" containerID="bf824fd498f08dcba70549c456ac18a0b2b1c882042f788014d8d6b1a385c577" Feb 03 09:12:03 crc kubenswrapper[4756]: I0203 09:12:03.154294 4756 scope.go:117] "RemoveContainer" containerID="d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b" Feb 03 09:12:03 crc kubenswrapper[4756]: E0203 09:12:03.154496 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-6srkq_openshift-multus(89e5d60f-f690-486f-b078-5ae9f98e1f3a)\"" pod="openshift-multus/multus-6srkq" podUID="89e5d60f-f690-486f-b078-5ae9f98e1f3a" Feb 03 09:12:04 crc kubenswrapper[4756]: I0203 09:12:04.158685 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/1.log" Feb 03 09:12:04 crc kubenswrapper[4756]: I0203 09:12:04.613555 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:04 crc kubenswrapper[4756]: I0203 09:12:04.613643 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:04 crc kubenswrapper[4756]: E0203 09:12:04.613693 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:04 crc kubenswrapper[4756]: I0203 09:12:04.613768 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:04 crc kubenswrapper[4756]: E0203 09:12:04.613813 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:04 crc kubenswrapper[4756]: E0203 09:12:04.613915 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:04 crc kubenswrapper[4756]: I0203 09:12:04.613952 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:04 crc kubenswrapper[4756]: E0203 09:12:04.614020 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:06 crc kubenswrapper[4756]: I0203 09:12:06.613744 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:06 crc kubenswrapper[4756]: E0203 09:12:06.614307 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:06 crc kubenswrapper[4756]: I0203 09:12:06.613821 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:06 crc kubenswrapper[4756]: I0203 09:12:06.613795 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:06 crc kubenswrapper[4756]: E0203 09:12:06.614378 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:06 crc kubenswrapper[4756]: I0203 09:12:06.613810 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:06 crc kubenswrapper[4756]: E0203 09:12:06.614501 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:06 crc kubenswrapper[4756]: E0203 09:12:06.614578 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:07 crc kubenswrapper[4756]: I0203 09:12:07.614216 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.173195 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/3.log" Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.176218 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerStarted","Data":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.176739 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.207245 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podStartSLOduration=100.207174703 podStartE2EDuration="1m40.207174703s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:08.203361808 +0000 UTC m=+119.353829193" watchObservedRunningTime="2026-02-03 09:12:08.207174703 +0000 UTC m=+119.357642148" Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.419643 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-k6pzt"] Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.419901 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:08 crc kubenswrapper[4756]: E0203 09:12:08.420103 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.613086 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.613148 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:08 crc kubenswrapper[4756]: E0203 09:12:08.613537 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:08 crc kubenswrapper[4756]: I0203 09:12:08.613243 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:08 crc kubenswrapper[4756]: E0203 09:12:08.613670 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:08 crc kubenswrapper[4756]: E0203 09:12:08.613822 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:09 crc kubenswrapper[4756]: E0203 09:12:09.618650 4756 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 03 09:12:09 crc kubenswrapper[4756]: E0203 09:12:09.721866 4756 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 03 09:12:10 crc kubenswrapper[4756]: I0203 09:12:10.613393 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:10 crc kubenswrapper[4756]: I0203 09:12:10.613391 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:10 crc kubenswrapper[4756]: E0203 09:12:10.613577 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:10 crc kubenswrapper[4756]: I0203 09:12:10.613437 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:10 crc kubenswrapper[4756]: E0203 09:12:10.613627 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:10 crc kubenswrapper[4756]: I0203 09:12:10.613405 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:10 crc kubenswrapper[4756]: E0203 09:12:10.613661 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:10 crc kubenswrapper[4756]: E0203 09:12:10.613700 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:12 crc kubenswrapper[4756]: I0203 09:12:12.613504 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:12 crc kubenswrapper[4756]: I0203 09:12:12.613557 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:12 crc kubenswrapper[4756]: E0203 09:12:12.613623 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:12 crc kubenswrapper[4756]: E0203 09:12:12.613694 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:12 crc kubenswrapper[4756]: I0203 09:12:12.613756 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:12 crc kubenswrapper[4756]: E0203 09:12:12.613808 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:12 crc kubenswrapper[4756]: I0203 09:12:12.613842 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:12 crc kubenswrapper[4756]: E0203 09:12:12.613889 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:14 crc kubenswrapper[4756]: I0203 09:12:14.613421 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:14 crc kubenswrapper[4756]: I0203 09:12:14.613525 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:14 crc kubenswrapper[4756]: I0203 09:12:14.613427 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:14 crc kubenswrapper[4756]: I0203 09:12:14.613418 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:14 crc kubenswrapper[4756]: E0203 09:12:14.613618 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:14 crc kubenswrapper[4756]: E0203 09:12:14.613746 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:14 crc kubenswrapper[4756]: E0203 09:12:14.613854 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:14 crc kubenswrapper[4756]: E0203 09:12:14.613942 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:14 crc kubenswrapper[4756]: E0203 09:12:14.722987 4756 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 03 09:12:15 crc kubenswrapper[4756]: I0203 09:12:15.613926 4756 scope.go:117] "RemoveContainer" containerID="d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b" Feb 03 09:12:16 crc kubenswrapper[4756]: I0203 09:12:16.211912 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/1.log" Feb 03 09:12:16 crc kubenswrapper[4756]: I0203 09:12:16.212229 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6srkq" event={"ID":"89e5d60f-f690-486f-b078-5ae9f98e1f3a","Type":"ContainerStarted","Data":"b0a499eb23a8e69d38e737205e6ed33a99d4a789663122520887ad03b0c3fb57"} Feb 03 09:12:16 crc kubenswrapper[4756]: I0203 09:12:16.613861 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:16 crc kubenswrapper[4756]: I0203 09:12:16.613886 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:16 crc kubenswrapper[4756]: I0203 09:12:16.613886 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:16 crc kubenswrapper[4756]: E0203 09:12:16.614006 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:16 crc kubenswrapper[4756]: I0203 09:12:16.614025 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:16 crc kubenswrapper[4756]: E0203 09:12:16.614109 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:16 crc kubenswrapper[4756]: E0203 09:12:16.614165 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:16 crc kubenswrapper[4756]: E0203 09:12:16.614202 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:18 crc kubenswrapper[4756]: I0203 09:12:18.613079 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:18 crc kubenswrapper[4756]: I0203 09:12:18.613146 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:18 crc kubenswrapper[4756]: I0203 09:12:18.613092 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:18 crc kubenswrapper[4756]: E0203 09:12:18.613242 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 03 09:12:18 crc kubenswrapper[4756]: I0203 09:12:18.613412 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:18 crc kubenswrapper[4756]: E0203 09:12:18.613635 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 03 09:12:18 crc kubenswrapper[4756]: E0203 09:12:18.613986 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 03 09:12:18 crc kubenswrapper[4756]: E0203 09:12:18.614187 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k6pzt" podUID="717b7410-dd64-44cb-ba9b-3436d82ebb95" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.613562 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.613572 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.613589 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.613728 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.616334 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.617859 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.617947 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.618115 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.618605 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 03 09:12:20 crc kubenswrapper[4756]: I0203 09:12:20.618703 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.307657 4756 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.459436 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bk8nb"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.459955 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.460717 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.461251 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.462252 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.463477 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.463623 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.468196 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9ghwh"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.470598 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.470885 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.471116 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.471421 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.471741 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.471974 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.472115 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.472255 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.472394 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.477685 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.478093 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.477698 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.478634 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.477747 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.479167 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.479192 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xn68m"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.479483 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.479718 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.479868 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.479967 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.480177 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.480638 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.486503 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dfj8c"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.486929 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.487324 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.487704 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.488694 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-6qrwb"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.489203 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.490194 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.490502 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.493511 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-7nst7"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.494123 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7nst7" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.494532 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lht77"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.494922 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.495045 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8jkfm"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.495617 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.500512 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nzcq8"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.501059 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.501494 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.501979 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.502891 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxxsv"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.503303 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.506530 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.506977 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.507223 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-66w8b"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.507510 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.507822 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.507965 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.511894 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.512363 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513385 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513531 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513557 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513576 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513703 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513772 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513845 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513912 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513948 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.513970 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514007 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514052 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514054 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514124 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514166 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514191 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514199 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514264 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514279 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514361 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514399 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514459 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514471 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514545 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514553 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514584 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514621 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514691 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514725 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514778 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514804 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514814 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514895 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514910 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514959 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514986 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515005 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514901 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515106 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515169 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515216 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515273 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515314 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515321 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515343 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515360 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515546 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.528977 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.529056 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.529276 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.529408 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.529525 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.529658 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.529777 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.529924 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.515550 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.530087 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.530235 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.530321 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.530399 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.514362 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.530816 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.530855 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.531379 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.532018 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.532197 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.532647 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.532958 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.533111 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.533437 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.533559 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.537082 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.538882 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.540624 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.540922 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.546325 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.547407 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.547689 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.548191 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.548358 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.569063 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.569361 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570244 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-client\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570297 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-serving-cert\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570291 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570320 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxgrt\" (UniqueName: \"kubernetes.io/projected/e1efc604-5ef7-445c-ac40-0e8cb2165c54-kube-api-access-gxgrt\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570340 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1efc604-5ef7-445c-ac40-0e8cb2165c54-serving-cert\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570355 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1efc604-5ef7-445c-ac40-0e8cb2165c54-trusted-ca\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570370 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-oauth-config\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570384 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-trusted-ca-bundle\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570399 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjg8l\" (UniqueName: \"kubernetes.io/projected/2c9fbee0-09f2-4796-a20b-683820300874-kube-api-access-cjg8l\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570417 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4cg5\" (UniqueName: \"kubernetes.io/projected/4dad28d0-454d-434b-b111-137c215aea4a-kube-api-access-d4cg5\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570452 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95gwc\" (UniqueName: \"kubernetes.io/projected/d984609f-8888-48c2-ba65-391882cdea2f-kube-api-access-95gwc\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570467 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-console-config\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570481 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-service-ca\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570495 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-oauth-serving-cert\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570532 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/067d7478-51bd-4d6f-af53-431c04d0eedd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qbnkf\" (UID: \"067d7478-51bd-4d6f-af53-431c04d0eedd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570467 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570559 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-audit-policies\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570579 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-config\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570611 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c9fbee0-09f2-4796-a20b-683820300874-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570627 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-ca\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570650 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d984609f-8888-48c2-ba65-391882cdea2f-audit-dir\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570670 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1efc604-5ef7-445c-ac40-0e8cb2165c54-config\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570686 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/864b3199-5d5c-407f-b8b7-dddfbe46bb78-serving-cert\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570803 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-encryption-config\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570852 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97754419-a9ec-4d4b-88be-72518ac8ef7d-config\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570881 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29gd2\" (UniqueName: \"kubernetes.io/projected/97754419-a9ec-4d4b-88be-72518ac8ef7d-kube-api-access-29gd2\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570901 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-service-ca\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570918 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-serving-cert\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570932 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/97754419-a9ec-4d4b-88be-72518ac8ef7d-machine-approver-tls\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570946 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570960 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570975 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-images\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.570989 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mvqr\" (UniqueName: \"kubernetes.io/projected/864b3199-5d5c-407f-b8b7-dddfbe46bb78-kube-api-access-7mvqr\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.571004 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr9zl\" (UniqueName: \"kubernetes.io/projected/067d7478-51bd-4d6f-af53-431c04d0eedd-kube-api-access-mr9zl\") pod \"cluster-samples-operator-665b6dd947-qbnkf\" (UID: \"067d7478-51bd-4d6f-af53-431c04d0eedd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.571018 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-etcd-client\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.571031 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-config\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.571093 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22zxs\" (UniqueName: \"kubernetes.io/projected/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-kube-api-access-22zxs\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.571121 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/97754419-a9ec-4d4b-88be-72518ac8ef7d-auth-proxy-config\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.571151 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.571171 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c9fbee0-09f2-4796-a20b-683820300874-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.574390 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.575200 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.577204 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.578026 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.579516 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.580007 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xn68m"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.580036 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.580008 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.580052 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.582789 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.584559 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.584795 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jwhwp"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.584835 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.585360 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.585409 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.585597 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.589949 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.590433 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.591594 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.597344 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.602992 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.604836 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bltft"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.608670 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.608745 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.608682 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.611481 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.611676 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.611404 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8jhbb"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.611869 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.617744 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6q85j"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.621669 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.621739 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.622088 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.622330 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.622587 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.623184 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.629812 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.635312 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.636083 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zgdp2"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.637076 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.641210 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.642169 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.642248 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.643421 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.645602 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.645792 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9ghwh"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.647671 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.660943 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jwhwp"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.663506 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.664660 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.664827 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nzcq8"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.666983 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dfj8c"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.668026 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7nst7"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.669290 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671564 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29gd2\" (UniqueName: \"kubernetes.io/projected/97754419-a9ec-4d4b-88be-72518ac8ef7d-kube-api-access-29gd2\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671590 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-service-ca\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671609 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-serving-cert\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671625 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/97754419-a9ec-4d4b-88be-72518ac8ef7d-machine-approver-tls\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671641 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671657 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671674 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-images\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671688 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mvqr\" (UniqueName: \"kubernetes.io/projected/864b3199-5d5c-407f-b8b7-dddfbe46bb78-kube-api-access-7mvqr\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671703 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr9zl\" (UniqueName: \"kubernetes.io/projected/067d7478-51bd-4d6f-af53-431c04d0eedd-kube-api-access-mr9zl\") pod \"cluster-samples-operator-665b6dd947-qbnkf\" (UID: \"067d7478-51bd-4d6f-af53-431c04d0eedd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671716 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-etcd-client\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671730 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-config\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671746 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22zxs\" (UniqueName: \"kubernetes.io/projected/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-kube-api-access-22zxs\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671764 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/97754419-a9ec-4d4b-88be-72518ac8ef7d-auth-proxy-config\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671786 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671802 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c9fbee0-09f2-4796-a20b-683820300874-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671818 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-client\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671837 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-serving-cert\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671853 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxgrt\" (UniqueName: \"kubernetes.io/projected/e1efc604-5ef7-445c-ac40-0e8cb2165c54-kube-api-access-gxgrt\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671868 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1efc604-5ef7-445c-ac40-0e8cb2165c54-serving-cert\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671881 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1efc604-5ef7-445c-ac40-0e8cb2165c54-trusted-ca\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671895 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-oauth-config\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671911 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-trusted-ca-bundle\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671925 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjg8l\" (UniqueName: \"kubernetes.io/projected/2c9fbee0-09f2-4796-a20b-683820300874-kube-api-access-cjg8l\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671941 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4cg5\" (UniqueName: \"kubernetes.io/projected/4dad28d0-454d-434b-b111-137c215aea4a-kube-api-access-d4cg5\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671962 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95gwc\" (UniqueName: \"kubernetes.io/projected/d984609f-8888-48c2-ba65-391882cdea2f-kube-api-access-95gwc\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671976 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-console-config\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.671989 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-service-ca\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.672004 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-oauth-serving-cert\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.672018 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/067d7478-51bd-4d6f-af53-431c04d0eedd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qbnkf\" (UID: \"067d7478-51bd-4d6f-af53-431c04d0eedd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.672036 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-audit-policies\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680885 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-config\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680926 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c9fbee0-09f2-4796-a20b-683820300874-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680961 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-ca\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680980 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d984609f-8888-48c2-ba65-391882cdea2f-audit-dir\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680998 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1efc604-5ef7-445c-ac40-0e8cb2165c54-config\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.681014 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/864b3199-5d5c-407f-b8b7-dddfbe46bb78-serving-cert\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.681085 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-encryption-config\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.681125 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97754419-a9ec-4d4b-88be-72518ac8ef7d-config\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.681149 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-oauth-config\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.679601 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/97754419-a9ec-4d4b-88be-72518ac8ef7d-machine-approver-tls\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.679803 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/067d7478-51bd-4d6f-af53-431c04d0eedd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qbnkf\" (UID: \"067d7478-51bd-4d6f-af53-431c04d0eedd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.679986 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.681222 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d984609f-8888-48c2-ba65-391882cdea2f-audit-dir\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680163 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-images\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680774 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c9fbee0-09f2-4796-a20b-683820300874-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.676373 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-audit-policies\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.679714 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1efc604-5ef7-445c-ac40-0e8cb2165c54-trusted-ca\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.681205 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-trusted-ca-bundle\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680389 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-client\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.677079 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-service-ca\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.675977 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/97754419-a9ec-4d4b-88be-72518ac8ef7d-auth-proxy-config\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.680788 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-etcd-client\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.676562 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.679167 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-config\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.678592 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-console-config\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.672326 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-service-ca\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.675082 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d984609f-8888-48c2-ba65-391882cdea2f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.681720 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c9fbee0-09f2-4796-a20b-683820300874-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.681936 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-etcd-ca\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.682304 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97754419-a9ec-4d4b-88be-72518ac8ef7d-config\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.682333 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.682870 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-serving-cert\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.683175 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1efc604-5ef7-445c-ac40-0e8cb2165c54-config\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.683311 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/864b3199-5d5c-407f-b8b7-dddfbe46bb78-config\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.683768 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-serving-cert\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.683806 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-oauth-serving-cert\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.687811 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d984609f-8888-48c2-ba65-391882cdea2f-encryption-config\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.687997 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1efc604-5ef7-445c-ac40-0e8cb2165c54-serving-cert\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.688431 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.689716 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.690684 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lht77"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.691695 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.692777 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxxsv"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.693705 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-ctwl6"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.694498 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.694667 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.695739 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.696675 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.697943 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/864b3199-5d5c-407f-b8b7-dddfbe46bb78-serving-cert\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.698026 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.698858 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8jkfm"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.704107 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.705154 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.705688 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.706766 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.708725 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zgdp2"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.712141 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.713931 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bltft"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.717211 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6qrwb"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.718894 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.720186 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bk8nb"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.721565 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-h2f8n"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.723054 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-h2f8n" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.723473 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.724210 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8jhbb"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.725318 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ctwl6"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.726384 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.727853 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-h2f8n"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.728862 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6q85j"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.730114 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.730787 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.731770 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sprfw"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.733693 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-qljx2"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.733855 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.736431 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.742823 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.744328 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sprfw"] Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.761485 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.783342 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.801983 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.822301 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.841759 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.862556 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.883019 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.902088 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.922074 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.941876 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.962743 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 03 09:12:27 crc kubenswrapper[4756]: I0203 09:12:27.982148 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.042302 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.062426 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.082258 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.102086 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.122265 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.143478 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.162952 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.183040 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.202310 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.222461 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.242554 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.262984 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.283270 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.302960 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.323129 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.341644 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.361215 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.381736 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.402965 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.422140 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.442508 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.462435 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.481635 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.502922 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.522757 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.542594 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.562814 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.582742 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.601675 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.621210 4756 request.go:700] Waited for 1.008715215s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-metrics&limit=500&resourceVersion=0 Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.623125 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.643121 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.670518 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.682557 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.702522 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.723227 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.759431 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.762268 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.782760 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.808384 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.821956 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.842893 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.862572 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.882485 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.902885 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.922866 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.942429 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.970164 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 03 09:12:28 crc kubenswrapper[4756]: I0203 09:12:28.982324 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.001794 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.023003 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.041418 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.062139 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.081884 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.101871 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.122021 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.142605 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.163160 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.181665 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.203062 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.223343 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.242338 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.278840 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29gd2\" (UniqueName: \"kubernetes.io/projected/97754419-a9ec-4d4b-88be-72518ac8ef7d-kube-api-access-29gd2\") pod \"machine-approver-56656f9798-f4xxl\" (UID: \"97754419-a9ec-4d4b-88be-72518ac8ef7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.308910 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr9zl\" (UniqueName: \"kubernetes.io/projected/067d7478-51bd-4d6f-af53-431c04d0eedd-kube-api-access-mr9zl\") pod \"cluster-samples-operator-665b6dd947-qbnkf\" (UID: \"067d7478-51bd-4d6f-af53-431c04d0eedd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.309047 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.328629 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22zxs\" (UniqueName: \"kubernetes.io/projected/f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e-kube-api-access-22zxs\") pod \"machine-api-operator-5694c8668f-bk8nb\" (UID: \"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.341122 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjg8l\" (UniqueName: \"kubernetes.io/projected/2c9fbee0-09f2-4796-a20b-683820300874-kube-api-access-cjg8l\") pod \"openshift-controller-manager-operator-756b6f6bc6-v76p8\" (UID: \"2c9fbee0-09f2-4796-a20b-683820300874\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.360573 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mvqr\" (UniqueName: \"kubernetes.io/projected/864b3199-5d5c-407f-b8b7-dddfbe46bb78-kube-api-access-7mvqr\") pod \"etcd-operator-b45778765-lht77\" (UID: \"864b3199-5d5c-407f-b8b7-dddfbe46bb78\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.386300 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxgrt\" (UniqueName: \"kubernetes.io/projected/e1efc604-5ef7-445c-ac40-0e8cb2165c54-kube-api-access-gxgrt\") pod \"console-operator-58897d9998-nzcq8\" (UID: \"e1efc604-5ef7-445c-ac40-0e8cb2165c54\") " pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.406086 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4cg5\" (UniqueName: \"kubernetes.io/projected/4dad28d0-454d-434b-b111-137c215aea4a-kube-api-access-d4cg5\") pod \"console-f9d7485db-6qrwb\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.422582 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.426601 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95gwc\" (UniqueName: \"kubernetes.io/projected/d984609f-8888-48c2-ba65-391882cdea2f-kube-api-access-95gwc\") pod \"apiserver-7bbb656c7d-wgxqm\" (UID: \"d984609f-8888-48c2-ba65-391882cdea2f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.437028 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.442417 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 03 09:12:29 crc kubenswrapper[4756]: W0203 09:12:29.455935 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97754419_a9ec_4d4b_88be_72518ac8ef7d.slice/crio-443c21827cbdf1e853449bb068e736be8d365d5b684db5faea94100e32fc2cf8 WatchSource:0}: Error finding container 443c21827cbdf1e853449bb068e736be8d365d5b684db5faea94100e32fc2cf8: Status 404 returned error can't find the container with id 443c21827cbdf1e853449bb068e736be8d365d5b684db5faea94100e32fc2cf8 Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.462663 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.465361 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.473582 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.482524 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.502160 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.522813 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.524408 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.537792 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.538972 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf"] Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.541909 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.577283 4756 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.581270 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.584194 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.604225 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.622309 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.640395 4756 request.go:700] Waited for 1.903321274s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.642572 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.661993 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.669075 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6qrwb"] Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.732375 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-trusted-ca\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.733386 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-tls\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.733477 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e579b706-5cc4-4a92-94f2-b22ad2135ffc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.733521 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.733550 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-certificates\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.733565 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-bound-sa-token\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.733578 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtdsg\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-kube-api-access-dtdsg\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.733604 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e579b706-5cc4-4a92-94f2-b22ad2135ffc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.734270 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:29 crc kubenswrapper[4756]: E0203 09:12:29.735041 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:30.235027839 +0000 UTC m=+141.385495214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.834829 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.834991 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e579b706-5cc4-4a92-94f2-b22ad2135ffc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835031 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835061 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-stats-auth\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835087 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfn78\" (UniqueName: \"kubernetes.io/projected/35e10745-2b1b-4843-a815-98bb6f85ca27-kube-api-access-qfn78\") pod \"dns-operator-744455d44c-8jkfm\" (UID: \"35e10745-2b1b-4843-a815-98bb6f85ca27\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835109 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlh8n\" (UniqueName: \"kubernetes.io/projected/268a0d65-02ea-4f46-8179-37115427a80d-kube-api-access-zlh8n\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835134 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhchl\" (UniqueName: \"kubernetes.io/projected/af06c306-9cb1-4716-8f23-73cc2f771af2-kube-api-access-zhchl\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835157 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-client-ca\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835190 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835208 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835245 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtdsg\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-kube-api-access-dtdsg\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835269 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c681fb4f-4fac-40de-b779-c7d9d57cf22c-metrics-tls\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835289 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-audit\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835312 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-certificates\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835335 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-bound-sa-token\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835365 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e579b706-5cc4-4a92-94f2-b22ad2135ffc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: E0203 09:12:29.835398 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:30.335370113 +0000 UTC m=+141.485837518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835479 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-config\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835616 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-metrics-certs\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835646 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-config\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835673 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-etcd-client\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835716 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835808 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e579b706-5cc4-4a92-94f2-b22ad2135ffc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835814 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-image-import-ca\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.835990 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d948f\" (UniqueName: \"kubernetes.io/projected/c4668221-e119-43fa-a7b4-ac39240893a3-kube-api-access-d948f\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.836239 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2wws\" (UniqueName: \"kubernetes.io/projected/625ab6f6-0677-42a9-871d-5f935d03a9b7-kube-api-access-c2wws\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.836957 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-certificates\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837546 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-service-ca-bundle\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837621 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/268a0d65-02ea-4f46-8179-37115427a80d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837663 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rnf7\" (UniqueName: \"kubernetes.io/projected/d61f6dad-2b6f-475c-8516-127db0c852c2-kube-api-access-2rnf7\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837690 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lbrr\" (UniqueName: \"kubernetes.io/projected/142cdcb7-ccff-4a63-99c5-e857b8eaa4df-kube-api-access-7lbrr\") pod \"downloads-7954f5f757-7nst7\" (UID: \"142cdcb7-ccff-4a63-99c5-e857b8eaa4df\") " pod="openshift-console/downloads-7954f5f757-7nst7" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837715 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed59c14-315c-43be-842e-f09920e0b18f-config\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837736 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4668221-e119-43fa-a7b4-ac39240893a3-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837788 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-serving-cert\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837808 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d61f6dad-2b6f-475c-8516-127db0c852c2-serving-cert\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837851 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wvm7\" (UniqueName: \"kubernetes.io/projected/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-kube-api-access-4wvm7\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837881 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.837902 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m92wh\" (UniqueName: \"kubernetes.io/projected/c681fb4f-4fac-40de-b779-c7d9d57cf22c-kube-api-access-m92wh\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.838084 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-etcd-serving-ca\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.838162 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/268a0d65-02ea-4f46-8179-37115427a80d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.838553 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lht77"] Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.838641 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-default-certificate\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.838682 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-serving-cert\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.838747 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-trusted-ca\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.838778 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eed59c14-315c-43be-842e-f09920e0b18f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.838910 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f848653-564d-4a41-ab3b-3e509a42076e-serving-cert\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839306 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-config\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839335 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-encryption-config\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839355 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-config\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839375 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c681fb4f-4fac-40de-b779-c7d9d57cf22c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839394 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1820186d-e8a3-4f09-a6db-01b7bbaa583c-node-pullsecrets\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839427 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c681fb4f-4fac-40de-b779-c7d9d57cf22c-trusted-ca\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839470 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt5dh\" (UniqueName: \"kubernetes.io/projected/3f848653-564d-4a41-ab3b-3e509a42076e-kube-api-access-jt5dh\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839495 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4668221-e119-43fa-a7b4-ac39240893a3-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839515 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839549 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1820186d-e8a3-4f09-a6db-01b7bbaa583c-audit-dir\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839567 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-client-ca\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.839973 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-tls\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840001 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/35e10745-2b1b-4843-a815-98bb6f85ca27-metrics-tls\") pod \"dns-operator-744455d44c-8jkfm\" (UID: \"35e10745-2b1b-4843-a815-98bb6f85ca27\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840046 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-trusted-ca\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840277 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eed59c14-315c-43be-842e-f09920e0b18f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840376 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af06c306-9cb1-4716-8f23-73cc2f771af2-serving-cert\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840482 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/268a0d65-02ea-4f46-8179-37115427a80d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840633 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/625ab6f6-0677-42a9-871d-5f935d03a9b7-service-ca-bundle\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840652 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-config\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840691 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hptxk\" (UniqueName: \"kubernetes.io/projected/1820186d-e8a3-4f09-a6db-01b7bbaa583c-kube-api-access-hptxk\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.840921 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e579b706-5cc4-4a92-94f2-b22ad2135ffc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.844917 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-tls\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: W0203 09:12:29.857775 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod864b3199_5d5c_407f_b8b7_dddfbe46bb78.slice/crio-ab2b05c8ec196e852bf61c64e8a9386b8cb8bbddda10fd061fb7b0a9055c6f40 WatchSource:0}: Error finding container ab2b05c8ec196e852bf61c64e8a9386b8cb8bbddda10fd061fb7b0a9055c6f40: Status 404 returned error can't find the container with id ab2b05c8ec196e852bf61c64e8a9386b8cb8bbddda10fd061fb7b0a9055c6f40 Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.875713 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bk8nb"] Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.881694 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtdsg\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-kube-api-access-dtdsg\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: W0203 09:12:29.892972 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf98e53dc_7bca_4a0f_a614_2ffbf3efdf9e.slice/crio-1917073358ef12ab7ee24b5d040517f19fbaaef7cca87df98c8788bee15c4d49 WatchSource:0}: Error finding container 1917073358ef12ab7ee24b5d040517f19fbaaef7cca87df98c8788bee15c4d49: Status 404 returned error can't find the container with id 1917073358ef12ab7ee24b5d040517f19fbaaef7cca87df98c8788bee15c4d49 Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.899370 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-bound-sa-token\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.905486 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8"] Feb 03 09:12:29 crc kubenswrapper[4756]: W0203 09:12:29.933065 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c9fbee0_09f2_4796_a20b_683820300874.slice/crio-b8f5e4c2ff3bcb1820d937079d330b9c70254221ac8c1bb5a892b365e539ffea WatchSource:0}: Error finding container b8f5e4c2ff3bcb1820d937079d330b9c70254221ac8c1bb5a892b365e539ffea: Status 404 returned error can't find the container with id b8f5e4c2ff3bcb1820d937079d330b9c70254221ac8c1bb5a892b365e539ffea Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.939485 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm"] Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941202 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj2gf\" (UniqueName: \"kubernetes.io/projected/6f020886-dc7a-49db-ad89-976ec3f8d771-kube-api-access-zj2gf\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941229 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gswt\" (UniqueName: \"kubernetes.io/projected/cf559b99-2353-49cd-8a10-2cd1d33a31f4-kube-api-access-9gswt\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941247 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3f642171-10cc-4c2d-b596-52b418d20ee8-metrics-tls\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941288 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1cbf7ea-2adc-4b59-a731-ee978309766d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941304 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-dir\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941369 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9tz6\" (UniqueName: \"kubernetes.io/projected/e742c048-9564-4b75-986b-a34639494020-kube-api-access-s9tz6\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941407 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941469 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cmml\" (UniqueName: \"kubernetes.io/projected/1daff10d-ff13-4c5a-9433-d33ade5b6486-kube-api-access-5cmml\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941531 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/888496f6-a8fd-4f24-a301-e9c1d883d844-node-bootstrap-token\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941622 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941642 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e5e0d3e-9130-4986-9369-71439f93c3ac-webhook-cert\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941677 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e742c048-9564-4b75-986b-a34639494020-config-volume\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941723 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrtsv\" (UniqueName: \"kubernetes.io/projected/3e5e0d3e-9130-4986-9369-71439f93c3ac-kube-api-access-jrtsv\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941761 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941781 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-serving-cert\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941840 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-mountpoint-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941902 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-client-ca\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941929 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf559b99-2353-49cd-8a10-2cd1d33a31f4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.941979 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942005 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5tfr\" (UniqueName: \"kubernetes.io/projected/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-kube-api-access-t5tfr\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942056 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c681fb4f-4fac-40de-b779-c7d9d57cf22c-metrics-tls\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942088 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/888496f6-a8fd-4f24-a301-e9c1d883d844-certs\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942237 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-config\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942281 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-config\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942319 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-etcd-client\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942345 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942359 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-available-featuregates\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942371 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aff499c5-85af-435b-b101-59bee677e283-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942399 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-policies\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942835 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.942978 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-config\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943048 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943069 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-image-import-ca\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943124 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d948f\" (UniqueName: \"kubernetes.io/projected/c4668221-e119-43fa-a7b4-ac39240893a3-kube-api-access-d948f\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943255 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-client-ca\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943377 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6e87ebde-e93e-4cbe-a868-8f09775fd15c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9sbkd\" (UID: \"6e87ebde-e93e-4cbe-a868-8f09775fd15c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943649 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96dmf\" (UniqueName: \"kubernetes.io/projected/6e87ebde-e93e-4cbe-a868-8f09775fd15c-kube-api-access-96dmf\") pod \"package-server-manager-789f6589d5-9sbkd\" (UID: \"6e87ebde-e93e-4cbe-a868-8f09775fd15c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943755 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-service-ca-bundle\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943756 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-image-import-ca\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943802 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e5e0d3e-9130-4986-9369-71439f93c3ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943852 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.943957 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/268a0d65-02ea-4f46-8179-37115427a80d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944009 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f35270c7-ed05-4b4e-bbb4-ad87efb8f940-cert\") pod \"ingress-canary-h2f8n\" (UID: \"f35270c7-ed05-4b4e-bbb4-ad87efb8f940\") " pod="openshift-ingress-canary/ingress-canary-h2f8n" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944033 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-signing-key\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944060 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-config\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944065 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f642171-10cc-4c2d-b596-52b418d20ee8-config-volume\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944426 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4668221-e119-43fa-a7b4-ac39240893a3-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944489 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lbrr\" (UniqueName: \"kubernetes.io/projected/142cdcb7-ccff-4a63-99c5-e857b8eaa4df-kube-api-access-7lbrr\") pod \"downloads-7954f5f757-7nst7\" (UID: \"142cdcb7-ccff-4a63-99c5-e857b8eaa4df\") " pod="openshift-console/downloads-7954f5f757-7nst7" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944528 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d61f6dad-2b6f-475c-8516-127db0c852c2-serving-cert\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944553 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944567 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-service-ca-bundle\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944585 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2qlg\" (UniqueName: \"kubernetes.io/projected/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-kube-api-access-h2qlg\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944606 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.944675 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/268a0d65-02ea-4f46-8179-37115427a80d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945072 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-default-certificate\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945131 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-serving-cert\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945159 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f1cbf7ea-2adc-4b59-a731-ee978309766d-images\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945356 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945493 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-encryption-config\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945552 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4668221-e119-43fa-a7b4-ac39240893a3-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945666 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c681fb4f-4fac-40de-b779-c7d9d57cf22c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945723 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-config\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945752 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1cbf7ea-2adc-4b59-a731-ee978309766d-proxy-tls\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945889 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945943 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.945978 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4668221-e119-43fa-a7b4-ac39240893a3-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946003 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946025 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1820186d-e8a3-4f09-a6db-01b7bbaa583c-audit-dir\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946050 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-client-ca\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946073 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946120 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/35e10745-2b1b-4843-a815-98bb6f85ca27-metrics-tls\") pod \"dns-operator-744455d44c-8jkfm\" (UID: \"35e10745-2b1b-4843-a815-98bb6f85ca27\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946147 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvcmz\" (UniqueName: \"kubernetes.io/projected/aff499c5-85af-435b-b101-59bee677e283-kube-api-access-xvcmz\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946175 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/268a0d65-02ea-4f46-8179-37115427a80d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946204 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-config\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946235 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eed59c14-315c-43be-842e-f09920e0b18f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946262 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hptxk\" (UniqueName: \"kubernetes.io/projected/1820186d-e8a3-4f09-a6db-01b7bbaa583c-kube-api-access-hptxk\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946289 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbj7x\" (UniqueName: \"kubernetes.io/projected/641583b7-728e-4e81-ad79-efcc524c5a10-kube-api-access-lbj7x\") pod \"control-plane-machine-set-operator-78cbb6b69f-qzff8\" (UID: \"641583b7-728e-4e81-ad79-efcc524c5a10\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946318 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-profile-collector-cert\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946348 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-stats-auth\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946384 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/268a0d65-02ea-4f46-8179-37115427a80d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946393 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-config\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946420 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm48v\" (UniqueName: \"kubernetes.io/projected/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-kube-api-access-fm48v\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946479 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfn78\" (UniqueName: \"kubernetes.io/projected/35e10745-2b1b-4843-a815-98bb6f85ca27-kube-api-access-qfn78\") pod \"dns-operator-744455d44c-8jkfm\" (UID: \"35e10745-2b1b-4843-a815-98bb6f85ca27\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946509 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf559b99-2353-49cd-8a10-2cd1d33a31f4-srv-cert\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946535 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlh8n\" (UniqueName: \"kubernetes.io/projected/268a0d65-02ea-4f46-8179-37115427a80d-kube-api-access-zlh8n\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946570 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946594 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhchl\" (UniqueName: \"kubernetes.io/projected/af06c306-9cb1-4716-8f23-73cc2f771af2-kube-api-access-zhchl\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946611 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1820186d-e8a3-4f09-a6db-01b7bbaa583c-audit-dir\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946627 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-plugins-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946662 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946725 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-audit\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946753 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-signing-cabundle\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946784 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946807 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946834 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-metrics-certs\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946855 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qzjk\" (UniqueName: \"kubernetes.io/projected/888496f6-a8fd-4f24-a301-e9c1d883d844-kube-api-access-9qzjk\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946910 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfq48\" (UniqueName: \"kubernetes.io/projected/f35270c7-ed05-4b4e-bbb4-ad87efb8f940-kube-api-access-zfq48\") pod \"ingress-canary-h2f8n\" (UID: \"f35270c7-ed05-4b4e-bbb4-ad87efb8f940\") " pod="openshift-ingress-canary/ingress-canary-h2f8n" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946936 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2wws\" (UniqueName: \"kubernetes.io/projected/625ab6f6-0677-42a9-871d-5f935d03a9b7-kube-api-access-c2wws\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946958 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.946986 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947013 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed59c14-315c-43be-842e-f09920e0b18f-config\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947037 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-serving-cert\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947060 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rnf7\" (UniqueName: \"kubernetes.io/projected/d61f6dad-2b6f-475c-8516-127db0c852c2-kube-api-access-2rnf7\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947081 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aff499c5-85af-435b-b101-59bee677e283-proxy-tls\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947084 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947105 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wvm7\" (UniqueName: \"kubernetes.io/projected/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-kube-api-access-4wvm7\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947129 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-srv-cert\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947155 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947182 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m92wh\" (UniqueName: \"kubernetes.io/projected/c681fb4f-4fac-40de-b779-c7d9d57cf22c-kube-api-access-m92wh\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947207 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-config\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947234 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rlqw\" (UniqueName: \"kubernetes.io/projected/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-kube-api-access-2rlqw\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947299 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-csi-data-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947622 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-config\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.947830 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-config\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: E0203 09:12:29.948781 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:30.448764304 +0000 UTC m=+141.599231759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.949944 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-audit\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.950276 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-etcd-client\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.950662 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d61f6dad-2b6f-475c-8516-127db0c852c2-serving-cert\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.950795 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/35e10745-2b1b-4843-a815-98bb6f85ca27-metrics-tls\") pod \"dns-operator-744455d44c-8jkfm\" (UID: \"35e10745-2b1b-4843-a815-98bb6f85ca27\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.951250 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c681fb4f-4fac-40de-b779-c7d9d57cf22c-metrics-tls\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.951608 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-serving-cert\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.951856 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-etcd-serving-ca\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952214 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9scz\" (UniqueName: \"kubernetes.io/projected/c818aba6-e4ed-442e-97c1-b73f5c99e404-kube-api-access-z9scz\") pod \"migrator-59844c95c7-dtkjs\" (UID: \"c818aba6-e4ed-442e-97c1-b73f5c99e404\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952265 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e5e0d3e-9130-4986-9369-71439f93c3ac-tmpfs\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952355 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f848653-564d-4a41-ab3b-3e509a42076e-serving-cert\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952400 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e742c048-9564-4b75-986b-a34639494020-secret-volume\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952706 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhdl6\" (UniqueName: \"kubernetes.io/projected/f1cbf7ea-2adc-4b59-a731-ee978309766d-kube-api-access-jhdl6\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952798 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eed59c14-315c-43be-842e-f09920e0b18f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952846 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-config\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952872 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-registration-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952917 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78qft\" (UniqueName: \"kubernetes.io/projected/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-kube-api-access-78qft\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.952950 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.953641 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af06c306-9cb1-4716-8f23-73cc2f771af2-config\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.953742 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1820186d-e8a3-4f09-a6db-01b7bbaa583c-node-pullsecrets\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.953788 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1820186d-e8a3-4f09-a6db-01b7bbaa583c-node-pullsecrets\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.953828 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4af6f100-b621-407d-9096-1c10d304db9c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jwhwp\" (UID: \"4af6f100-b621-407d-9096-1c10d304db9c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.953864 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c681fb4f-4fac-40de-b779-c7d9d57cf22c-trusted-ca\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.953900 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt5dh\" (UniqueName: \"kubernetes.io/projected/3f848653-564d-4a41-ab3b-3e509a42076e-kube-api-access-jt5dh\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.953964 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1820186d-e8a3-4f09-a6db-01b7bbaa583c-etcd-serving-ca\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954090 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-socket-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954263 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954592 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/641583b7-728e-4e81-ad79-efcc524c5a10-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qzff8\" (UID: \"641583b7-728e-4e81-ad79-efcc524c5a10\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954725 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954824 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cwd2\" (UniqueName: \"kubernetes.io/projected/4af6f100-b621-407d-9096-1c10d304db9c-kube-api-access-7cwd2\") pod \"multus-admission-controller-857f4d67dd-jwhwp\" (UID: \"4af6f100-b621-407d-9096-1c10d304db9c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954928 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7jcw\" (UniqueName: \"kubernetes.io/projected/3f642171-10cc-4c2d-b596-52b418d20ee8-kube-api-access-z7jcw\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.955032 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af06c306-9cb1-4716-8f23-73cc2f771af2-serving-cert\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.955128 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/625ab6f6-0677-42a9-871d-5f935d03a9b7-service-ca-bundle\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954767 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed59c14-315c-43be-842e-f09920e0b18f-config\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954614 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-client-ca\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.955309 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-serving-cert\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954931 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c681fb4f-4fac-40de-b779-c7d9d57cf22c-trusted-ca\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.955535 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-default-certificate\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.954959 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.955596 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-stats-auth\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.955868 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1820186d-e8a3-4f09-a6db-01b7bbaa583c-encryption-config\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.956106 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eed59c14-315c-43be-842e-f09920e0b18f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.956750 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/625ab6f6-0677-42a9-871d-5f935d03a9b7-service-ca-bundle\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.959104 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4668221-e119-43fa-a7b4-ac39240893a3-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.960116 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.960221 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f848653-564d-4a41-ab3b-3e509a42076e-serving-cert\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.960262 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/625ab6f6-0677-42a9-871d-5f935d03a9b7-metrics-certs\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.966645 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af06c306-9cb1-4716-8f23-73cc2f771af2-serving-cert\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.967393 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/268a0d65-02ea-4f46-8179-37115427a80d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.974362 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nzcq8"] Feb 03 09:12:29 crc kubenswrapper[4756]: I0203 09:12:29.977730 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6cb0296-3bd3-4a27-9a43-0b346fcfbd20-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pr6ct\" (UID: \"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:29 crc kubenswrapper[4756]: W0203 09:12:29.984621 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1efc604_5ef7_445c_ac40_0e8cb2165c54.slice/crio-013f64f0511ce2fdfed3ba9b83d64eb499317353a5836c485dc242854c4c32dd WatchSource:0}: Error finding container 013f64f0511ce2fdfed3ba9b83d64eb499317353a5836c485dc242854c4c32dd: Status 404 returned error can't find the container with id 013f64f0511ce2fdfed3ba9b83d64eb499317353a5836c485dc242854c4c32dd Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.000620 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d948f\" (UniqueName: \"kubernetes.io/projected/c4668221-e119-43fa-a7b4-ac39240893a3-kube-api-access-d948f\") pod \"kube-storage-version-migrator-operator-b67b599dd-42dt5\" (UID: \"c4668221-e119-43fa-a7b4-ac39240893a3\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.019721 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/268a0d65-02ea-4f46-8179-37115427a80d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.036605 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lbrr\" (UniqueName: \"kubernetes.io/projected/142cdcb7-ccff-4a63-99c5-e857b8eaa4df-kube-api-access-7lbrr\") pod \"downloads-7954f5f757-7nst7\" (UID: \"142cdcb7-ccff-4a63-99c5-e857b8eaa4df\") " pod="openshift-console/downloads-7954f5f757-7nst7" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.057760 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hptxk\" (UniqueName: \"kubernetes.io/projected/1820186d-e8a3-4f09-a6db-01b7bbaa583c-kube-api-access-hptxk\") pod \"apiserver-76f77b778f-9ghwh\" (UID: \"1820186d-e8a3-4f09-a6db-01b7bbaa583c\") " pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.060774 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061007 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfq48\" (UniqueName: \"kubernetes.io/projected/f35270c7-ed05-4b4e-bbb4-ad87efb8f940-kube-api-access-zfq48\") pod \"ingress-canary-h2f8n\" (UID: \"f35270c7-ed05-4b4e-bbb4-ad87efb8f940\") " pod="openshift-ingress-canary/ingress-canary-h2f8n" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061045 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061071 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061098 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aff499c5-85af-435b-b101-59bee677e283-proxy-tls\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061124 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-srv-cert\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061149 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-config\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061169 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rlqw\" (UniqueName: \"kubernetes.io/projected/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-kube-api-access-2rlqw\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061196 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-csi-data-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061225 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9scz\" (UniqueName: \"kubernetes.io/projected/c818aba6-e4ed-442e-97c1-b73f5c99e404-kube-api-access-z9scz\") pod \"migrator-59844c95c7-dtkjs\" (UID: \"c818aba6-e4ed-442e-97c1-b73f5c99e404\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061245 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e5e0d3e-9130-4986-9369-71439f93c3ac-tmpfs\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061266 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e742c048-9564-4b75-986b-a34639494020-secret-volume\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061287 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhdl6\" (UniqueName: \"kubernetes.io/projected/f1cbf7ea-2adc-4b59-a731-ee978309766d-kube-api-access-jhdl6\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061314 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-registration-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061336 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78qft\" (UniqueName: \"kubernetes.io/projected/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-kube-api-access-78qft\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061360 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061384 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4af6f100-b621-407d-9096-1c10d304db9c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jwhwp\" (UID: \"4af6f100-b621-407d-9096-1c10d304db9c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061410 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-socket-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061432 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061468 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/641583b7-728e-4e81-ad79-efcc524c5a10-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qzff8\" (UID: \"641583b7-728e-4e81-ad79-efcc524c5a10\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061491 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061511 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cwd2\" (UniqueName: \"kubernetes.io/projected/4af6f100-b621-407d-9096-1c10d304db9c-kube-api-access-7cwd2\") pod \"multus-admission-controller-857f4d67dd-jwhwp\" (UID: \"4af6f100-b621-407d-9096-1c10d304db9c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061533 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7jcw\" (UniqueName: \"kubernetes.io/projected/3f642171-10cc-4c2d-b596-52b418d20ee8-kube-api-access-z7jcw\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061555 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj2gf\" (UniqueName: \"kubernetes.io/projected/6f020886-dc7a-49db-ad89-976ec3f8d771-kube-api-access-zj2gf\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061577 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gswt\" (UniqueName: \"kubernetes.io/projected/cf559b99-2353-49cd-8a10-2cd1d33a31f4-kube-api-access-9gswt\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061595 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3f642171-10cc-4c2d-b596-52b418d20ee8-metrics-tls\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061613 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1cbf7ea-2adc-4b59-a731-ee978309766d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061634 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-dir\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061656 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9tz6\" (UniqueName: \"kubernetes.io/projected/e742c048-9564-4b75-986b-a34639494020-kube-api-access-s9tz6\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061676 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061696 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cmml\" (UniqueName: \"kubernetes.io/projected/1daff10d-ff13-4c5a-9433-d33ade5b6486-kube-api-access-5cmml\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061717 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/888496f6-a8fd-4f24-a301-e9c1d883d844-node-bootstrap-token\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061737 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e5e0d3e-9130-4986-9369-71439f93c3ac-webhook-cert\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061758 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e742c048-9564-4b75-986b-a34639494020-config-volume\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061779 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrtsv\" (UniqueName: \"kubernetes.io/projected/3e5e0d3e-9130-4986-9369-71439f93c3ac-kube-api-access-jrtsv\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061801 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061823 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-serving-cert\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061845 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-mountpoint-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061871 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf559b99-2353-49cd-8a10-2cd1d33a31f4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061894 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5tfr\" (UniqueName: \"kubernetes.io/projected/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-kube-api-access-t5tfr\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061914 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/888496f6-a8fd-4f24-a301-e9c1d883d844-certs\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061946 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061967 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aff499c5-85af-435b-b101-59bee677e283-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.061986 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-policies\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062011 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6e87ebde-e93e-4cbe-a868-8f09775fd15c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9sbkd\" (UID: \"6e87ebde-e93e-4cbe-a868-8f09775fd15c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062041 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96dmf\" (UniqueName: \"kubernetes.io/projected/6e87ebde-e93e-4cbe-a868-8f09775fd15c-kube-api-access-96dmf\") pod \"package-server-manager-789f6589d5-9sbkd\" (UID: \"6e87ebde-e93e-4cbe-a868-8f09775fd15c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062279 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e5e0d3e-9130-4986-9369-71439f93c3ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062302 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062322 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f35270c7-ed05-4b4e-bbb4-ad87efb8f940-cert\") pod \"ingress-canary-h2f8n\" (UID: \"f35270c7-ed05-4b4e-bbb4-ad87efb8f940\") " pod="openshift-ingress-canary/ingress-canary-h2f8n" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062342 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-signing-key\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062361 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f642171-10cc-4c2d-b596-52b418d20ee8-config-volume\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062383 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062404 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2qlg\" (UniqueName: \"kubernetes.io/projected/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-kube-api-access-h2qlg\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062425 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062462 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f1cbf7ea-2adc-4b59-a731-ee978309766d-images\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062486 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062517 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1cbf7ea-2adc-4b59-a731-ee978309766d-proxy-tls\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062539 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062559 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062582 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062604 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvcmz\" (UniqueName: \"kubernetes.io/projected/aff499c5-85af-435b-b101-59bee677e283-kube-api-access-xvcmz\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062629 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbj7x\" (UniqueName: \"kubernetes.io/projected/641583b7-728e-4e81-ad79-efcc524c5a10-kube-api-access-lbj7x\") pod \"control-plane-machine-set-operator-78cbb6b69f-qzff8\" (UID: \"641583b7-728e-4e81-ad79-efcc524c5a10\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062648 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-profile-collector-cert\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062671 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-config\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062692 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm48v\" (UniqueName: \"kubernetes.io/projected/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-kube-api-access-fm48v\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062719 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf559b99-2353-49cd-8a10-2cd1d33a31f4-srv-cert\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062758 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-plugins-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062790 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-signing-cabundle\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062810 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062834 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.062878 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qzjk\" (UniqueName: \"kubernetes.io/projected/888496f6-a8fd-4f24-a301-e9c1d883d844-kube-api-access-9qzjk\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.063118 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:30.563101732 +0000 UTC m=+141.713569107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.075123 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.075216 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-mountpoint-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.075602 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-dir\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.076613 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-registration-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.077243 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-socket-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.078687 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f1cbf7ea-2adc-4b59-a731-ee978309766d-images\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.078985 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-serving-cert\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.079766 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.079881 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1cbf7ea-2adc-4b59-a731-ee978309766d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.079939 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-policies\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.080302 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.081954 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.082069 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-plugins-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.082733 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f642171-10cc-4c2d-b596-52b418d20ee8-config-volume\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.082889 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aff499c5-85af-435b-b101-59bee677e283-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.083103 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-config\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.083467 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.083559 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.083755 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6f020886-dc7a-49db-ad89-976ec3f8d771-csi-data-dir\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.084134 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e5e0d3e-9130-4986-9369-71439f93c3ac-tmpfs\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.084349 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-signing-cabundle\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.084556 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-config\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.084638 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1cbf7ea-2adc-4b59-a731-ee978309766d-proxy-tls\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.086848 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.087714 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.088993 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/641583b7-728e-4e81-ad79-efcc524c5a10-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qzff8\" (UID: \"641583b7-728e-4e81-ad79-efcc524c5a10\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.089077 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4af6f100-b621-407d-9096-1c10d304db9c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jwhwp\" (UID: \"4af6f100-b621-407d-9096-1c10d304db9c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.089133 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/888496f6-a8fd-4f24-a301-e9c1d883d844-node-bootstrap-token\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.089553 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/888496f6-a8fd-4f24-a301-e9c1d883d844-certs\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.089609 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e742c048-9564-4b75-986b-a34639494020-config-volume\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.095046 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.095229 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.095301 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e5e0d3e-9130-4986-9369-71439f93c3ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.095392 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf559b99-2353-49cd-8a10-2cd1d33a31f4-srv-cert\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.095811 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.095864 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.095986 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e5e0d3e-9130-4986-9369-71439f93c3ac-webhook-cert\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096029 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e742c048-9564-4b75-986b-a34639494020-secret-volume\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096098 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6e87ebde-e93e-4cbe-a868-8f09775fd15c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9sbkd\" (UID: \"6e87ebde-e93e-4cbe-a868-8f09775fd15c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096119 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-srv-cert\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096417 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf559b99-2353-49cd-8a10-2cd1d33a31f4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096455 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3f642171-10cc-4c2d-b596-52b418d20ee8-metrics-tls\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096531 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096554 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f35270c7-ed05-4b4e-bbb4-ad87efb8f940-cert\") pod \"ingress-canary-h2f8n\" (UID: \"f35270c7-ed05-4b4e-bbb4-ad87efb8f940\") " pod="openshift-ingress-canary/ingress-canary-h2f8n" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096559 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-signing-key\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.096724 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c681fb4f-4fac-40de-b779-c7d9d57cf22c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.097239 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.098362 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aff499c5-85af-435b-b101-59bee677e283-proxy-tls\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.098699 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-profile-collector-cert\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.099034 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.099715 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2wws\" (UniqueName: \"kubernetes.io/projected/625ab6f6-0677-42a9-871d-5f935d03a9b7-kube-api-access-c2wws\") pod \"router-default-5444994796-66w8b\" (UID: \"625ab6f6-0677-42a9-871d-5f935d03a9b7\") " pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.105924 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.117434 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7nst7" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.123370 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m92wh\" (UniqueName: \"kubernetes.io/projected/c681fb4f-4fac-40de-b779-c7d9d57cf22c-kube-api-access-m92wh\") pod \"ingress-operator-5b745b69d9-8vwd4\" (UID: \"c681fb4f-4fac-40de-b779-c7d9d57cf22c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.142160 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wvm7\" (UniqueName: \"kubernetes.io/projected/c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed-kube-api-access-4wvm7\") pod \"openshift-config-operator-7777fb866f-2mtqs\" (UID: \"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.157374 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rnf7\" (UniqueName: \"kubernetes.io/projected/d61f6dad-2b6f-475c-8516-127db0c852c2-kube-api-access-2rnf7\") pod \"route-controller-manager-6576b87f9c-hzt8q\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.159937 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.163958 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.164548 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:30.664533507 +0000 UTC m=+141.815000882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.169537 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.174669 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.178964 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfn78\" (UniqueName: \"kubernetes.io/projected/35e10745-2b1b-4843-a815-98bb6f85ca27-kube-api-access-qfn78\") pod \"dns-operator-744455d44c-8jkfm\" (UID: \"35e10745-2b1b-4843-a815-98bb6f85ca27\") " pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.181074 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.202286 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlh8n\" (UniqueName: \"kubernetes.io/projected/268a0d65-02ea-4f46-8179-37115427a80d-kube-api-access-zlh8n\") pod \"cluster-image-registry-operator-dc59b4c8b-kwhqd\" (UID: \"268a0d65-02ea-4f46-8179-37115427a80d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.229000 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.248952 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhchl\" (UniqueName: \"kubernetes.io/projected/af06c306-9cb1-4716-8f23-73cc2f771af2-kube-api-access-zhchl\") pod \"authentication-operator-69f744f599-dfj8c\" (UID: \"af06c306-9cb1-4716-8f23-73cc2f771af2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.250396 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.259919 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eed59c14-315c-43be-842e-f09920e0b18f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r9s2p\" (UID: \"eed59c14-315c-43be-842e-f09920e0b18f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.264911 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.265731 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:30.765704912 +0000 UTC m=+141.916172287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.275065 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.279812 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6qrwb" event={"ID":"4dad28d0-454d-434b-b111-137c215aea4a","Type":"ContainerStarted","Data":"1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.279859 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6qrwb" event={"ID":"4dad28d0-454d-434b-b111-137c215aea4a","Type":"ContainerStarted","Data":"1eb0e66ccf3bd9b01f57ba9ef6d90a956da14150825527898a44842048b76028"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.280893 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-66w8b" event={"ID":"625ab6f6-0677-42a9-871d-5f935d03a9b7","Type":"ContainerStarted","Data":"f3cf88428a581a2f0dfd741379f94d2a231819bee96b555e78dc2573ea03f03e"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.284831 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" event={"ID":"2c9fbee0-09f2-4796-a20b-683820300874","Type":"ContainerStarted","Data":"5e9acb8f256bb58d0864f3f701a5adf753811fc598e835a9c8dec5ddf17f2d44"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.284877 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" event={"ID":"2c9fbee0-09f2-4796-a20b-683820300874","Type":"ContainerStarted","Data":"b8f5e4c2ff3bcb1820d937079d330b9c70254221ac8c1bb5a892b365e539ffea"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.311433 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7nst7"] Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.315391 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" event={"ID":"97754419-a9ec-4d4b-88be-72518ac8ef7d","Type":"ContainerStarted","Data":"068e55e5b3f1002858ae4ad162f08b635219c8fa2a3199f03ac8b03d8858220a"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.315462 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" event={"ID":"97754419-a9ec-4d4b-88be-72518ac8ef7d","Type":"ContainerStarted","Data":"3998eaa9877d36f362f99b0f9d7bb9a7e09040513791079f85e1e72a6d2082a7"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.315492 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" event={"ID":"97754419-a9ec-4d4b-88be-72518ac8ef7d","Type":"ContainerStarted","Data":"443c21827cbdf1e853449bb068e736be8d365d5b684db5faea94100e32fc2cf8"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.316338 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt5dh\" (UniqueName: \"kubernetes.io/projected/3f848653-564d-4a41-ab3b-3e509a42076e-kube-api-access-jt5dh\") pod \"controller-manager-879f6c89f-xn68m\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.316479 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qzjk\" (UniqueName: \"kubernetes.io/projected/888496f6-a8fd-4f24-a301-e9c1d883d844-kube-api-access-9qzjk\") pod \"machine-config-server-qljx2\" (UID: \"888496f6-a8fd-4f24-a301-e9c1d883d844\") " pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.325861 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" event={"ID":"067d7478-51bd-4d6f-af53-431c04d0eedd","Type":"ContainerStarted","Data":"da751200ceeae0521301a7828a7d718f66c3044e9701f4ba37050c2f9b49caf3"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.325908 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" event={"ID":"067d7478-51bd-4d6f-af53-431c04d0eedd","Type":"ContainerStarted","Data":"3dcc0abf3c6a44b451890e3641ffe02e2889dc1c8a5db5e2a980e747825ca463"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.325919 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" event={"ID":"067d7478-51bd-4d6f-af53-431c04d0eedd","Type":"ContainerStarted","Data":"029f0d788b4d65aad0588d16ec8e330b960b37d3ce14bf8a555af967864c00a3"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.328146 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" event={"ID":"d984609f-8888-48c2-ba65-391882cdea2f","Type":"ContainerStarted","Data":"8842d3958582b942f7d7fe98c7f9fa0049567cafa0a3c9069841c8f2bfffa1e6"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.329070 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" event={"ID":"e1efc604-5ef7-445c-ac40-0e8cb2165c54","Type":"ContainerStarted","Data":"1f2fdd4614dd3747ae313f5e2b341b65af71f4356374bb1a785624e1a3f073a7"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.329103 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" event={"ID":"e1efc604-5ef7-445c-ac40-0e8cb2165c54","Type":"ContainerStarted","Data":"013f64f0511ce2fdfed3ba9b83d64eb499317353a5836c485dc242854c4c32dd"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.329993 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.332932 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" event={"ID":"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e","Type":"ContainerStarted","Data":"22a8a1592bc3f376be404aaa289ba3397a4c0191f0cdb37674807f234517cc4c"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.332982 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" event={"ID":"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e","Type":"ContainerStarted","Data":"1917073358ef12ab7ee24b5d040517f19fbaaef7cca87df98c8788bee15c4d49"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.335839 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" event={"ID":"864b3199-5d5c-407f-b8b7-dddfbe46bb78","Type":"ContainerStarted","Data":"735f276fd3c0c77e6799377ee12f5368ab6c60fb9cf4e2ab4a7693551e215fee"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.335920 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" event={"ID":"864b3199-5d5c-407f-b8b7-dddfbe46bb78","Type":"ContainerStarted","Data":"ab2b05c8ec196e852bf61c64e8a9386b8cb8bbddda10fd061fb7b0a9055c6f40"} Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.336321 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfq48\" (UniqueName: \"kubernetes.io/projected/f35270c7-ed05-4b4e-bbb4-ad87efb8f940-kube-api-access-zfq48\") pod \"ingress-canary-h2f8n\" (UID: \"f35270c7-ed05-4b4e-bbb4-ad87efb8f940\") " pod="openshift-ingress-canary/ingress-canary-h2f8n" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.337465 4756 patch_prober.go:28] interesting pod/console-operator-58897d9998-nzcq8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.337509 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" podUID="e1efc604-5ef7-445c-ac40-0e8cb2165c54" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.345398 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj2gf\" (UniqueName: \"kubernetes.io/projected/6f020886-dc7a-49db-ad89-976ec3f8d771-kube-api-access-zj2gf\") pod \"csi-hostpathplugin-sprfw\" (UID: \"6f020886-dc7a-49db-ad89-976ec3f8d771\") " pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.352111 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-sprfw" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.356811 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.357131 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cwd2\" (UniqueName: \"kubernetes.io/projected/4af6f100-b621-407d-9096-1c10d304db9c-kube-api-access-7cwd2\") pod \"multus-admission-controller-857f4d67dd-jwhwp\" (UID: \"4af6f100-b621-407d-9096-1c10d304db9c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.358266 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qljx2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.366323 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.369135 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:30.868564113 +0000 UTC m=+142.019031488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.380208 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7jcw\" (UniqueName: \"kubernetes.io/projected/3f642171-10cc-4c2d-b596-52b418d20ee8-kube-api-access-z7jcw\") pod \"dns-default-ctwl6\" (UID: \"3f642171-10cc-4c2d-b596-52b418d20ee8\") " pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.408819 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9tz6\" (UniqueName: \"kubernetes.io/projected/e742c048-9564-4b75-986b-a34639494020-kube-api-access-s9tz6\") pod \"collect-profiles-29501820-8bb2z\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:30 crc kubenswrapper[4756]: W0203 09:12:30.416783 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod142cdcb7_ccff_4a63_99c5_e857b8eaa4df.slice/crio-390ba973ba1973be0ad56377253aa41b7936d43af63618c1fddb0e060d48bf7b WatchSource:0}: Error finding container 390ba973ba1973be0ad56377253aa41b7936d43af63618c1fddb0e060d48bf7b: Status 404 returned error can't find the container with id 390ba973ba1973be0ad56377253aa41b7936d43af63618c1fddb0e060d48bf7b Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.423211 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6a977b9-ab26-48bb-9dba-f49089aa3d4b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-42p25\" (UID: \"d6a977b9-ab26-48bb-9dba-f49089aa3d4b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.431636 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.439872 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cmml\" (UniqueName: \"kubernetes.io/projected/1daff10d-ff13-4c5a-9433-d33ade5b6486-kube-api-access-5cmml\") pod \"oauth-openshift-558db77b4-8jhbb\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.446717 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.466925 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhdl6\" (UniqueName: \"kubernetes.io/projected/f1cbf7ea-2adc-4b59-a731-ee978309766d-kube-api-access-jhdl6\") pod \"machine-config-operator-74547568cd-gsgxt\" (UID: \"f1cbf7ea-2adc-4b59-a731-ee978309766d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.468062 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.468749 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:30.96872812 +0000 UTC m=+142.119195495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.480477 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5tfr\" (UniqueName: \"kubernetes.io/projected/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-kube-api-access-t5tfr\") pod \"marketplace-operator-79b997595-bltft\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.492384 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.499630 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gswt\" (UniqueName: \"kubernetes.io/projected/cf559b99-2353-49cd-8a10-2cd1d33a31f4-kube-api-access-9gswt\") pod \"olm-operator-6b444d44fb-vdsfv\" (UID: \"cf559b99-2353-49cd-8a10-2cd1d33a31f4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.502696 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.520072 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78qft\" (UniqueName: \"kubernetes.io/projected/5eb78f90-4bc7-402f-96d6-ed8bbabb5660-kube-api-access-78qft\") pod \"openshift-apiserver-operator-796bbdcf4f-25jr4\" (UID: \"5eb78f90-4bc7-402f-96d6-ed8bbabb5660\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.522238 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.524988 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.549660 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.556197 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2qlg\" (UniqueName: \"kubernetes.io/projected/dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f-kube-api-access-h2qlg\") pod \"service-ca-9c57cc56f-zgdp2\" (UID: \"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f\") " pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.556392 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.570030 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.570330 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.070318571 +0000 UTC m=+142.220785946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.573063 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.581630 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.585125 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.585915 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96dmf\" (UniqueName: \"kubernetes.io/projected/6e87ebde-e93e-4cbe-a868-8f09775fd15c-kube-api-access-96dmf\") pod \"package-server-manager-789f6589d5-9sbkd\" (UID: \"6e87ebde-e93e-4cbe-a868-8f09775fd15c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.593227 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrtsv\" (UniqueName: \"kubernetes.io/projected/3e5e0d3e-9130-4986-9369-71439f93c3ac-kube-api-access-jrtsv\") pod \"packageserver-d55dfcdfc-8gmxs\" (UID: \"3e5e0d3e-9130-4986-9369-71439f93c3ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.598128 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm48v\" (UniqueName: \"kubernetes.io/projected/065d7db7-d4ca-4ca3-9cc3-d5abaeff069f-kube-api-access-fm48v\") pod \"service-ca-operator-777779d784-6q85j\" (UID: \"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.606044 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.638895 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.638899 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-h2f8n" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.644041 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.644089 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.645716 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rlqw\" (UniqueName: \"kubernetes.io/projected/789f68b4-cf41-4a3d-ac8f-4014c1bd81a7-kube-api-access-2rlqw\") pod \"catalog-operator-68c6474976-qz4nl\" (UID: \"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.660739 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvcmz\" (UniqueName: \"kubernetes.io/projected/aff499c5-85af-435b-b101-59bee677e283-kube-api-access-xvcmz\") pod \"machine-config-controller-84d6567774-sjcqf\" (UID: \"aff499c5-85af-435b-b101-59bee677e283\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.666184 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9scz\" (UniqueName: \"kubernetes.io/projected/c818aba6-e4ed-442e-97c1-b73f5c99e404-kube-api-access-z9scz\") pod \"migrator-59844c95c7-dtkjs\" (UID: \"c818aba6-e4ed-442e-97c1-b73f5c99e404\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.671744 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.671966 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.171947864 +0000 UTC m=+142.322415229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.672108 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.672668 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.172659262 +0000 UTC m=+142.323126637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.679238 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbj7x\" (UniqueName: \"kubernetes.io/projected/641583b7-728e-4e81-ad79-efcc524c5a10-kube-api-access-lbj7x\") pod \"control-plane-machine-set-operator-78cbb6b69f-qzff8\" (UID: \"641583b7-728e-4e81-ad79-efcc524c5a10\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.698407 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4"] Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.701278 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct"] Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.772573 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.772892 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.272876831 +0000 UTC m=+142.423344206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.795075 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.812139 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" Feb 03 09:12:30 crc kubenswrapper[4756]: W0203 09:12:30.813866 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc681fb4f_4fac_40de_b779_c7d9d57cf22c.slice/crio-61268b4780ae276601d72f4032b1e54acb3a024d52aa77a6693a09513e76e47f WatchSource:0}: Error finding container 61268b4780ae276601d72f4032b1e54acb3a024d52aa77a6693a09513e76e47f: Status 404 returned error can't find the container with id 61268b4780ae276601d72f4032b1e54acb3a024d52aa77a6693a09513e76e47f Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.842620 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.847529 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.849767 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q"] Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.867397 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.870734 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5"] Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.874280 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.874686 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.3746698 +0000 UTC m=+142.525137175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.895528 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.929073 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs"] Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.975117 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.975243 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.475216442 +0000 UTC m=+142.625683817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: I0203 09:12:30.975384 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:30 crc kubenswrapper[4756]: E0203 09:12:30.975768 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.475737002 +0000 UTC m=+142.626204377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:30 crc kubenswrapper[4756]: W0203 09:12:30.994343 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4668221_e119_43fa_a7b4_ac39240893a3.slice/crio-6fb1d8b97c98ea6700ff78f4ecc3ada992777c4fd83973d22ae47e62ccf9a944 WatchSource:0}: Error finding container 6fb1d8b97c98ea6700ff78f4ecc3ada992777c4fd83973d22ae47e62ccf9a944: Status 404 returned error can't find the container with id 6fb1d8b97c98ea6700ff78f4ecc3ada992777c4fd83973d22ae47e62ccf9a944 Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.076295 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.076634 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.576618446 +0000 UTC m=+142.727085811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.177757 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.178096 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.678082262 +0000 UTC m=+142.828549637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.278581 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.279194 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.779181525 +0000 UTC m=+142.929648900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.366329 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" event={"ID":"d61f6dad-2b6f-475c-8516-127db0c852c2","Type":"ContainerStarted","Data":"5a013ce0da93436a9256ffa45255470f734a9b1e895a504aec1decdd199d08a5"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.366599 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" event={"ID":"d61f6dad-2b6f-475c-8516-127db0c852c2","Type":"ContainerStarted","Data":"73348d585e9315b5978e593697163d9751f2a34edccf501897f236c347d7b6ca"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.396988 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" event={"ID":"c681fb4f-4fac-40de-b779-c7d9d57cf22c","Type":"ContainerStarted","Data":"51ad9c760ef86b4893e71056776276a1197f5eb6180188ad02a8244345b25595"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.397037 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" event={"ID":"c681fb4f-4fac-40de-b779-c7d9d57cf22c","Type":"ContainerStarted","Data":"61268b4780ae276601d72f4032b1e54acb3a024d52aa77a6693a09513e76e47f"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.399074 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.399596 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:31.899583364 +0000 UTC m=+143.050050739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.400846 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" event={"ID":"c4668221-e119-43fa-a7b4-ac39240893a3","Type":"ContainerStarted","Data":"6fb1d8b97c98ea6700ff78f4ecc3ada992777c4fd83973d22ae47e62ccf9a944"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.401538 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" event={"ID":"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed","Type":"ContainerStarted","Data":"fa1c4544d0be38f7dfbac2e3aa31d31da5baa7d6169eb1f01baa990041d410f9"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.402980 4756 generic.go:334] "Generic (PLEG): container finished" podID="d984609f-8888-48c2-ba65-391882cdea2f" containerID="be49f5b9c66e061fd32de202b792aa490ba0cd045ba313477396a13a998c2d50" exitCode=0 Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.403116 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" event={"ID":"d984609f-8888-48c2-ba65-391882cdea2f","Type":"ContainerDied","Data":"be49f5b9c66e061fd32de202b792aa490ba0cd045ba313477396a13a998c2d50"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.410823 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" event={"ID":"f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e","Type":"ContainerStarted","Data":"bfebb0430f6918aea307a86962fca2aef4295815bb4c0ab8aa2ff93fafc9e4e0"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.412158 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" event={"ID":"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20","Type":"ContainerStarted","Data":"f054253a692f3bfc243df4b18266ec75e59b53cc02500fcc68334bf2fe420310"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.414817 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-66w8b" event={"ID":"625ab6f6-0677-42a9-871d-5f935d03a9b7","Type":"ContainerStarted","Data":"5b8ab833dcb30c49f57d841129aecee9e261bb5e4bd4d4bcb46cdf830661b579"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.420281 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qljx2" event={"ID":"888496f6-a8fd-4f24-a301-e9c1d883d844","Type":"ContainerStarted","Data":"1139f533c02a7a3b56572bcbd1d15732319252b4c19ce4678ede2ce8dbe1dca6"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.420366 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qljx2" event={"ID":"888496f6-a8fd-4f24-a301-e9c1d883d844","Type":"ContainerStarted","Data":"9617a2abd587942938c5699909b4bad2cc83084662e512f9eac131596c099407"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.425956 4756 patch_prober.go:28] interesting pod/console-operator-58897d9998-nzcq8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.426002 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" podUID="e1efc604-5ef7-445c-ac40-0e8cb2165c54" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.427428 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7nst7" event={"ID":"142cdcb7-ccff-4a63-99c5-e857b8eaa4df","Type":"ContainerStarted","Data":"995d63cb32cbd933f055aa88d1fc17f268bae896f423207757bcb0f617832fa0"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.427491 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7nst7" event={"ID":"142cdcb7-ccff-4a63-99c5-e857b8eaa4df","Type":"ContainerStarted","Data":"390ba973ba1973be0ad56377253aa41b7936d43af63618c1fddb0e060d48bf7b"} Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.427868 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7nst7" Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.452720 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-v76p8" podStartSLOduration=123.452697038 podStartE2EDuration="2m3.452697038s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:31.45249938 +0000 UTC m=+142.602966755" watchObservedRunningTime="2026-02-03 09:12:31.452697038 +0000 UTC m=+142.603164413" Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.501014 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.502583 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.002559938 +0000 UTC m=+143.153027303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.534767 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-lht77" podStartSLOduration=123.534747155 podStartE2EDuration="2m3.534747155s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:31.533805669 +0000 UTC m=+142.684273054" watchObservedRunningTime="2026-02-03 09:12:31.534747155 +0000 UTC m=+142.685214530" Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.603473 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.603997 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.103966993 +0000 UTC m=+143.254434368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.658471 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" podStartSLOduration=123.658428009 podStartE2EDuration="2m3.658428009s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:31.65635619 +0000 UTC m=+142.806823585" watchObservedRunningTime="2026-02-03 09:12:31.658428009 +0000 UTC m=+142.808895384" Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.687198 4756 patch_prober.go:28] interesting pod/downloads-7954f5f757-7nst7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.687261 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7nst7" podUID="142cdcb7-ccff-4a63-99c5-e857b8eaa4df" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.705422 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.706146 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.206127546 +0000 UTC m=+143.356594921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.753402 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f4xxl" podStartSLOduration=123.753382707 podStartE2EDuration="2m3.753382707s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:31.751012057 +0000 UTC m=+142.901479422" watchObservedRunningTime="2026-02-03 09:12:31.753382707 +0000 UTC m=+142.903850082" Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.807335 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.807934 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.307922825 +0000 UTC m=+143.458390200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:31 crc kubenswrapper[4756]: I0203 09:12:31.911012 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:31 crc kubenswrapper[4756]: E0203 09:12:31.911399 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.411380398 +0000 UTC m=+143.561847773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.012623 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.013090 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.513073123 +0000 UTC m=+143.663540498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.030043 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9ghwh"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.037185 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dfj8c"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.083296 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sprfw"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.113955 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.114094 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.614065282 +0000 UTC m=+143.764532657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.114315 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.114740 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.614726048 +0000 UTC m=+143.765193423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.138367 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qbnkf" podStartSLOduration=124.138344717 podStartE2EDuration="2m4.138344717s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.136434715 +0000 UTC m=+143.286902090" watchObservedRunningTime="2026-02-03 09:12:32.138344717 +0000 UTC m=+143.288812102" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.160779 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.172687 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:32 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:32 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:32 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.172741 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.215945 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.216416 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.716395352 +0000 UTC m=+143.866862727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.262131 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-bk8nb" podStartSLOduration=123.262109184 podStartE2EDuration="2m3.262109184s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.260485382 +0000 UTC m=+143.410952757" watchObservedRunningTime="2026-02-03 09:12:32.262109184 +0000 UTC m=+143.412576559" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.316529 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-6qrwb" podStartSLOduration=124.316512878 podStartE2EDuration="2m4.316512878s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.302897989 +0000 UTC m=+143.453365364" watchObservedRunningTime="2026-02-03 09:12:32.316512878 +0000 UTC m=+143.466980253" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.324533 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.325355 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.326175 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.826163195 +0000 UTC m=+143.976630570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.333566 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-h2f8n"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.336305 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.360113 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z"] Feb 03 09:12:32 crc kubenswrapper[4756]: W0203 09:12:32.377783 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6a977b9_ab26_48bb_9dba_f49089aa3d4b.slice/crio-b6bb703434378f1c9a13a24fad6e95fdef6cf4f0be4d3c466d564b7cd3bce9be WatchSource:0}: Error finding container b6bb703434378f1c9a13a24fad6e95fdef6cf4f0be4d3c466d564b7cd3bce9be: Status 404 returned error can't find the container with id b6bb703434378f1c9a13a24fad6e95fdef6cf4f0be4d3c466d564b7cd3bce9be Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.406677 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.409285 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xn68m"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.410410 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.411851 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.425250 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ctwl6"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.426674 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.426912 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:32.926897014 +0000 UTC m=+144.077364389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.440816 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8jhbb"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.448796 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zgdp2"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.459998 4756 csr.go:261] certificate signing request csr-pmd54 is approved, waiting to be issued Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.468196 4756 csr.go:257] certificate signing request csr-pmd54 is issued Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.474088 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jwhwp"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.474176 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-8jkfm"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.480045 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.481977 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.490782 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.500018 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.500075 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bltft"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.506094 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.509541 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf"] Feb 03 09:12:32 crc kubenswrapper[4756]: W0203 09:12:32.524496 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5eb78f90_4bc7_402f_96d6_ed8bbabb5660.slice/crio-c6d6320688f9b3171862182e6883dcfecb4498c1bec76f75755464468903f81d WatchSource:0}: Error finding container c6d6320688f9b3171862182e6883dcfecb4498c1bec76f75755464468903f81d: Status 404 returned error can't find the container with id c6d6320688f9b3171862182e6883dcfecb4498c1bec76f75755464468903f81d Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.527737 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sprfw" event={"ID":"6f020886-dc7a-49db-ad89-976ec3f8d771","Type":"ContainerStarted","Data":"d3a1d1bf31397578d7d00f60f30577bc207b72680cb497d2e8ee2b95782e6b8b"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.528381 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.528762 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.028745595 +0000 UTC m=+144.179212970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.544395 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" event={"ID":"c4668221-e119-43fa-a7b4-ac39240893a3","Type":"ContainerStarted","Data":"6b0ceb713e276fcd3ef261127d73eee0db0bde8066ddeffd078b34056a2b507f"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.554351 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" event={"ID":"d984609f-8888-48c2-ba65-391882cdea2f","Type":"ContainerStarted","Data":"a88c0f6cb2517bfb28f9fb74cc4ece14feae07d314e896b76071436ff70332b6"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.557145 4756 generic.go:334] "Generic (PLEG): container finished" podID="c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed" containerID="98c75345b44364bb82ede719ddd24b0f9cc74565bfbb1e2c9b3f3feb143af993" exitCode=0 Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.557193 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" event={"ID":"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed","Type":"ContainerDied","Data":"98c75345b44364bb82ede719ddd24b0f9cc74565bfbb1e2c9b3f3feb143af993"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.557415 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-66w8b" podStartSLOduration=124.557393877 podStartE2EDuration="2m4.557393877s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.555085039 +0000 UTC m=+143.705552414" watchObservedRunningTime="2026-02-03 09:12:32.557393877 +0000 UTC m=+143.707861252" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.561176 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-h2f8n" event={"ID":"f35270c7-ed05-4b4e-bbb4-ad87efb8f940","Type":"ContainerStarted","Data":"735657b412c0a7b2aaff4f7eee751e007c757ac7bfc5cf32022e2d1c994ba37b"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.561976 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" event={"ID":"268a0d65-02ea-4f46-8179-37115427a80d","Type":"ContainerStarted","Data":"01cde95d03d0899a5fb1656af427db77c3c22874dec9e42a0523a8237f8a0de0"} Feb 03 09:12:32 crc kubenswrapper[4756]: W0203 09:12:32.563641 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaff499c5_85af_435b_b101_59bee677e283.slice/crio-4e8259692c69a69cfa5798ddef09a80328ca910c5a4957cb902cc3b4321fb37a WatchSource:0}: Error finding container 4e8259692c69a69cfa5798ddef09a80328ca910c5a4957cb902cc3b4321fb37a: Status 404 returned error can't find the container with id 4e8259692c69a69cfa5798ddef09a80328ca910c5a4957cb902cc3b4321fb37a Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.575759 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" event={"ID":"c681fb4f-4fac-40de-b779-c7d9d57cf22c","Type":"ContainerStarted","Data":"84769d213c923eeb1a01f3417d64d027d020822fce9b4441850a58e5b1b9d7b6"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.592882 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-7nst7" podStartSLOduration=124.592851498 podStartE2EDuration="2m4.592851498s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.588343467 +0000 UTC m=+143.738810852" watchObservedRunningTime="2026-02-03 09:12:32.592851498 +0000 UTC m=+143.743318863" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.630756 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.632015 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.13199989 +0000 UTC m=+144.282467265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.639653 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.675388 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6q85j"] Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.677871 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" event={"ID":"af06c306-9cb1-4716-8f23-73cc2f771af2","Type":"ContainerStarted","Data":"961d9db7023bdead916cf19f4a73baf3e94a6688378b42a3d560af11cdc98b53"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.677914 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" event={"ID":"af06c306-9cb1-4716-8f23-73cc2f771af2","Type":"ContainerStarted","Data":"00d1b7a05d0ee148fc9ea727631ad032550ee27a93566656d021b88e6a7089e6"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.681335 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-qljx2" podStartSLOduration=5.6813170490000005 podStartE2EDuration="5.681317049s" podCreationTimestamp="2026-02-03 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.641577376 +0000 UTC m=+143.792044751" watchObservedRunningTime="2026-02-03 09:12:32.681317049 +0000 UTC m=+143.831784424" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.693075 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" event={"ID":"f1cbf7ea-2adc-4b59-a731-ee978309766d","Type":"ContainerStarted","Data":"cbefa910de7326122381fe03c7c6f32d97ba0d485dd8a340e395f28400c0b4a6"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.695425 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-42dt5" podStartSLOduration=123.695409156 podStartE2EDuration="2m3.695409156s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.695048574 +0000 UTC m=+143.845515949" watchObservedRunningTime="2026-02-03 09:12:32.695409156 +0000 UTC m=+143.845876531" Feb 03 09:12:32 crc kubenswrapper[4756]: W0203 09:12:32.719394 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod065d7db7_d4ca_4ca3_9cc3_d5abaeff069f.slice/crio-ea099a1d74f146c91d713e9334826312b7726ae1dc6d0ce1a041d25dcc71655b WatchSource:0}: Error finding container ea099a1d74f146c91d713e9334826312b7726ae1dc6d0ce1a041d25dcc71655b: Status 404 returned error can't find the container with id ea099a1d74f146c91d713e9334826312b7726ae1dc6d0ce1a041d25dcc71655b Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.732793 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.734477 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.234464415 +0000 UTC m=+144.384931790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.749251 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" event={"ID":"c6cb0296-3bd3-4a27-9a43-0b346fcfbd20","Type":"ContainerStarted","Data":"5b8ed9f8e77336b77b997af9648bf6d8edd08cc48a9cb4645c684691ddb242f6"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.759900 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" event={"ID":"d6a977b9-ab26-48bb-9dba-f49089aa3d4b","Type":"ContainerStarted","Data":"b6bb703434378f1c9a13a24fad6e95fdef6cf4f0be4d3c466d564b7cd3bce9be"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.775808 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" event={"ID":"1820186d-e8a3-4f09-a6db-01b7bbaa583c","Type":"ContainerStarted","Data":"ab888df9d5669f9e75660b89c53d37ef9a885bfe5ab2df5aa9e89bb9247dee6d"} Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.776319 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.779627 4756 patch_prober.go:28] interesting pod/downloads-7954f5f757-7nst7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.779697 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7nst7" podUID="142cdcb7-ccff-4a63-99c5-e857b8eaa4df" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.797324 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8vwd4" podStartSLOduration=124.79730111 podStartE2EDuration="2m4.79730111s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.744772738 +0000 UTC m=+143.895240113" watchObservedRunningTime="2026-02-03 09:12:32.79730111 +0000 UTC m=+143.947768475" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.799859 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" podStartSLOduration=123.799852487 podStartE2EDuration="2m3.799852487s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.793185863 +0000 UTC m=+143.943653268" watchObservedRunningTime="2026-02-03 09:12:32.799852487 +0000 UTC m=+143.950319852" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.824238 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dfj8c" podStartSLOduration=124.824214695 podStartE2EDuration="2m4.824214695s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.816242981 +0000 UTC m=+143.966710356" watchObservedRunningTime="2026-02-03 09:12:32.824214695 +0000 UTC m=+143.974682070" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.842406 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.843266 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.343233861 +0000 UTC m=+144.493701236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.861540 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" podStartSLOduration=123.861521077 podStartE2EDuration="2m3.861521077s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.859992569 +0000 UTC m=+144.010459934" watchObservedRunningTime="2026-02-03 09:12:32.861521077 +0000 UTC m=+144.011988452" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.896632 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pr6ct" podStartSLOduration=124.896610875 podStartE2EDuration="2m4.896610875s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:32.896501461 +0000 UTC m=+144.046968836" watchObservedRunningTime="2026-02-03 09:12:32.896610875 +0000 UTC m=+144.047078250" Feb 03 09:12:32 crc kubenswrapper[4756]: I0203 09:12:32.954303 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:32 crc kubenswrapper[4756]: E0203 09:12:32.954628 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.454617085 +0000 UTC m=+144.605084450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.055021 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.055680 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.555662016 +0000 UTC m=+144.706129391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.155990 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.156367 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.656355703 +0000 UTC m=+144.806823078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.165162 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:33 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:33 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:33 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.165190 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.257015 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.257428 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.757413555 +0000 UTC m=+144.907880930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.358342 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.358988 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.858971874 +0000 UTC m=+145.009439249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.368565 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.460529 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.460894 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:33.960880598 +0000 UTC m=+145.111347973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.479566 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-03 09:07:32 +0000 UTC, rotation deadline is 2026-10-23 00:47:31.961188566 +0000 UTC Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.479607 4756 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6279h34m58.481584194s for next certificate rotation Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.562458 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.562927 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.062895176 +0000 UTC m=+145.213362601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.663118 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.663303 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.163271832 +0000 UTC m=+145.313739217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.663698 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.664160 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.164145284 +0000 UTC m=+145.314612669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.764726 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.765020 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.264992847 +0000 UTC m=+145.415460222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.765047 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.765437 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.265427684 +0000 UTC m=+145.415895059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.785831 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" event={"ID":"641583b7-728e-4e81-ad79-efcc524c5a10","Type":"ContainerStarted","Data":"4866295f2657a92e1e3eb9c95a00bf41c1d76eeafd01ea028304ad59101497ce"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.785891 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" event={"ID":"641583b7-728e-4e81-ad79-efcc524c5a10","Type":"ContainerStarted","Data":"45dc91319b89636b695c11d068fd24b6b9e679689957b24c4d59242b8a3f69c4"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.803666 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qzff8" podStartSLOduration=124.803650821 podStartE2EDuration="2m4.803650821s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:33.802091441 +0000 UTC m=+144.952558816" watchObservedRunningTime="2026-02-03 09:12:33.803650821 +0000 UTC m=+144.954118196" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.822197 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" event={"ID":"d6a977b9-ab26-48bb-9dba-f49089aa3d4b","Type":"ContainerStarted","Data":"05ba4d1bdaf3e393a6240f9da6490265f7e0c9605fa46efd0bd67a79bde275b1"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.824552 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" event={"ID":"e742c048-9564-4b75-986b-a34639494020","Type":"ContainerStarted","Data":"ff07b217cc643322909bb9ab3d7a77b96673d00b9ba816c5d024d1b3056c44a2"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.824615 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" event={"ID":"e742c048-9564-4b75-986b-a34639494020","Type":"ContainerStarted","Data":"e8f2ab8f5bb9faa51ec8e7a09f426a6bd186e6a9b166f43f6101f8104e86b746"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.839695 4756 generic.go:334] "Generic (PLEG): container finished" podID="1820186d-e8a3-4f09-a6db-01b7bbaa583c" containerID="661eddaed7e063e5865b330b6acbeafee54292ebcceb72a2ec2890b7b5395065" exitCode=0 Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.839789 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" event={"ID":"1820186d-e8a3-4f09-a6db-01b7bbaa583c","Type":"ContainerDied","Data":"661eddaed7e063e5865b330b6acbeafee54292ebcceb72a2ec2890b7b5395065"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.843513 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-42p25" podStartSLOduration=124.843499539 podStartE2EDuration="2m4.843499539s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:33.842257093 +0000 UTC m=+144.992724488" watchObservedRunningTime="2026-02-03 09:12:33.843499539 +0000 UTC m=+144.993966914" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.847646 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" event={"ID":"6e87ebde-e93e-4cbe-a868-8f09775fd15c","Type":"ContainerStarted","Data":"adc024c319fe2e25368fb8a82dd3afc4718ef20b157349820d38b5851645bc23"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.847692 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" event={"ID":"6e87ebde-e93e-4cbe-a868-8f09775fd15c","Type":"ContainerStarted","Data":"232b2c5b70a9fec869f1dbab95a4581fd5bd62707cf83a7ead762c5b937e2386"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.857044 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" event={"ID":"3f848653-564d-4a41-ab3b-3e509a42076e","Type":"ContainerStarted","Data":"86c4adbb56c70b5e3b7cbab7f5522d2218131fe50218672931f616f916548e18"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.857139 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" event={"ID":"3f848653-564d-4a41-ab3b-3e509a42076e","Type":"ContainerStarted","Data":"5470dcebe391f9bbfbada4d47fc184f66ee17e2875571918a072348d6b0ed4e8"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.860830 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.865806 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" event={"ID":"cf559b99-2353-49cd-8a10-2cd1d33a31f4","Type":"ContainerStarted","Data":"d36f02713acb5d620b7c7d544ae2a965a41f8c3f87598405269c7c656bf69711"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.865854 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" event={"ID":"cf559b99-2353-49cd-8a10-2cd1d33a31f4","Type":"ContainerStarted","Data":"2ef4d6250c7ed6ee8767acb5f64759f7e733df6da1fc0f64f2abe05d995cc99f"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.866217 4756 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xn68m container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.866273 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" podUID="3f848653-564d-4a41-ab3b-3e509a42076e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.866344 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.866483 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.866815 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.366796017 +0000 UTC m=+145.517263392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.867154 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.867545 4756 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vdsfv container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.867577 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" podUID="cf559b99-2353-49cd-8a10-2cd1d33a31f4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.868239 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.368225331 +0000 UTC m=+145.518692766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.874747 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" podStartSLOduration=124.87472814 podStartE2EDuration="2m4.87472814s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:33.873327907 +0000 UTC m=+145.023795282" watchObservedRunningTime="2026-02-03 09:12:33.87472814 +0000 UTC m=+145.025195515" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.887028 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" event={"ID":"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7","Type":"ContainerStarted","Data":"d6a0ea6c215fc767b62c0508027679e2016987c6c9f592809f44ba169ab3db8d"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.887070 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" event={"ID":"789f68b4-cf41-4a3d-ac8f-4014c1bd81a7","Type":"ContainerStarted","Data":"7a02c3ad0f1f1d0f912afc8ad4b1bf744c72055a61ebae8a113e5cfd411cb969"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.887549 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.891082 4756 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qz4nl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.891137 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" podUID="789f68b4-cf41-4a3d-ac8f-4014c1bd81a7" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.924111 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" event={"ID":"eed59c14-315c-43be-842e-f09920e0b18f","Type":"ContainerStarted","Data":"ce5d6b455e6650b3fdc04f7e81a0eaa79a8558b38013c0eebf63f4f382689a84"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.924167 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" event={"ID":"eed59c14-315c-43be-842e-f09920e0b18f","Type":"ContainerStarted","Data":"d8214fb018c716ece17261df2382ee3d64b22a302e66d0c6fabe5d484a00638c"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.937323 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" event={"ID":"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7","Type":"ContainerStarted","Data":"ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.937656 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" event={"ID":"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7","Type":"ContainerStarted","Data":"03765366fd902afb2e3296b98dcda116ed0ca540ee2a743629c001158c87cb39"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.937673 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.941076 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" event={"ID":"c818aba6-e4ed-442e-97c1-b73f5c99e404","Type":"ContainerStarted","Data":"01f16bd7635b1c67ff71d39ae39f45bf9723d94ac70da06f2ca7b3c686c23568"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.941128 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" event={"ID":"c818aba6-e4ed-442e-97c1-b73f5c99e404","Type":"ContainerStarted","Data":"f53a6c498ae5a44fafea6e5009e6cfef06171b8cd857097051707c55512414e7"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.942873 4756 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bltft container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.942929 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" podUID="8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.958041 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" podStartSLOduration=124.958027414 podStartE2EDuration="2m4.958027414s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:33.931970241 +0000 UTC m=+145.082437616" watchObservedRunningTime="2026-02-03 09:12:33.958027414 +0000 UTC m=+145.108494789" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.959454 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" podStartSLOduration=125.959435388 podStartE2EDuration="2m5.959435388s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:33.957210903 +0000 UTC m=+145.107678278" watchObservedRunningTime="2026-02-03 09:12:33.959435388 +0000 UTC m=+145.109902763" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.964161 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" event={"ID":"4af6f100-b621-407d-9096-1c10d304db9c","Type":"ContainerStarted","Data":"35ee90649616861b3d9f303e1405daa95a7625619ef194c5c9876c9c1863d526"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.964213 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" event={"ID":"4af6f100-b621-407d-9096-1c10d304db9c","Type":"ContainerStarted","Data":"0a6134c071742df9016ee31d89127b434a83d84ecf37c86de02cbf1f5b1e4fce"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.968723 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" event={"ID":"c6b2655b-b2d5-4ec5-89b8-71c2b7c713ed","Type":"ContainerStarted","Data":"695af2da534769898f918820fcbe35c74adddfcc03c15d6e3b8dad2834d06a43"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.969192 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.969693 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:33 crc kubenswrapper[4756]: E0203 09:12:33.970454 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.470427637 +0000 UTC m=+145.620895012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.984979 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" event={"ID":"aff499c5-85af-435b-b101-59bee677e283","Type":"ContainerStarted","Data":"c263155cc5cb0c22e695568a9999b9fdb0e4ee4932bc90d2660fe26d94c49a09"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.985037 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" event={"ID":"aff499c5-85af-435b-b101-59bee677e283","Type":"ContainerStarted","Data":"bc769a9188208c0ce03434d9f6411a5d646a0002aaf276865c27b9d862ac662f"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.985050 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" event={"ID":"aff499c5-85af-435b-b101-59bee677e283","Type":"ContainerStarted","Data":"4e8259692c69a69cfa5798ddef09a80328ca910c5a4957cb902cc3b4321fb37a"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.987129 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" podStartSLOduration=124.987117643 podStartE2EDuration="2m4.987117643s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:33.986159716 +0000 UTC m=+145.136627091" watchObservedRunningTime="2026-02-03 09:12:33.987117643 +0000 UTC m=+145.137585028" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.991590 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ctwl6" event={"ID":"3f642171-10cc-4c2d-b596-52b418d20ee8","Type":"ContainerStarted","Data":"3a7e74d06c1f6ba05d62181a677ceba711398b31dba33cc1265f31ec98225f7b"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.991638 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ctwl6" event={"ID":"3f642171-10cc-4c2d-b596-52b418d20ee8","Type":"ContainerStarted","Data":"3c86ae39ecffbb2ec01a2a6451d8e1cc5bd997a685afb294318b7adf9b40feac"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.994682 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" event={"ID":"1daff10d-ff13-4c5a-9433-d33ade5b6486","Type":"ContainerStarted","Data":"c0bfc447ce13e025405a965e87ec319c7452d5914c5b5a1a8c9f29c76680feab"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.994729 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" event={"ID":"1daff10d-ff13-4c5a-9433-d33ade5b6486","Type":"ContainerStarted","Data":"22008310cf66c57e6bd80fcaae6182b187163b39c18451e288da0e760f94159d"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.995780 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.999624 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" event={"ID":"3e5e0d3e-9130-4986-9369-71439f93c3ac","Type":"ContainerStarted","Data":"6e1b6fa8ddbbc5ede64dd5e1e2ebe37fbec7fca87495466c58317f4fa8ba5430"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.999676 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" event={"ID":"3e5e0d3e-9130-4986-9369-71439f93c3ac","Type":"ContainerStarted","Data":"2137cbc0f2e23edb6db0d70480ff2c0bf88353109e725295bfd3234bf31cd21c"} Feb 03 09:12:33 crc kubenswrapper[4756]: I0203 09:12:33.999919 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.002690 4756 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8jhbb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.40:6443/healthz\": dial tcp 10.217.0.40:6443: connect: connection refused" start-of-body= Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.002728 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" podUID="1daff10d-ff13-4c5a-9433-d33ade5b6486" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.40:6443/healthz\": dial tcp 10.217.0.40:6443: connect: connection refused" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.002934 4756 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8gmxs container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.002965 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" podUID="3e5e0d3e-9130-4986-9369-71439f93c3ac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.009755 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" podStartSLOduration=125.009740035 podStartE2EDuration="2m5.009740035s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.009238446 +0000 UTC m=+145.159705821" watchObservedRunningTime="2026-02-03 09:12:34.009740035 +0000 UTC m=+145.160207410" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.039029 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sprfw" event={"ID":"6f020886-dc7a-49db-ad89-976ec3f8d771","Type":"ContainerStarted","Data":"401e5d9039beca9d5202b490c072b8b8909bdd640520bc52a71e22868ad671ce"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.042302 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" event={"ID":"35e10745-2b1b-4843-a815-98bb6f85ca27","Type":"ContainerStarted","Data":"7db0c20a66c55f88db89203c0f60984742690397ac08447dc8b5959bd3b1e798"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.042360 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" event={"ID":"35e10745-2b1b-4843-a815-98bb6f85ca27","Type":"ContainerStarted","Data":"a1a01ff9f5f63974654b11d59a931cf03d114e689a3790dd4b3280e0e03ea04d"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.047941 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" event={"ID":"268a0d65-02ea-4f46-8179-37115427a80d","Type":"ContainerStarted","Data":"a539fa200718aa4f4ac5088bd1ce820f379cee174fedcc0f4667de01c954fb98"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.051412 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" event={"ID":"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f","Type":"ContainerStarted","Data":"ea71be63e467e59ce1ca3b1837015151e8e34a699fd194c0d7ae68fa1c2f1f29"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.051461 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" event={"ID":"dd0561f0-55aa-4885-9cb8-c32b0e7e0a7f","Type":"ContainerStarted","Data":"63ce537b3190acdc74d605003bb483aa8074f62ae39affd8593d4657bdb2c23f"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.056110 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" event={"ID":"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f","Type":"ContainerStarted","Data":"33e0417d5ef049558ebc31c21c693d6665e635b8eb3d97c3b12bb734888c306f"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.056155 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" event={"ID":"065d7db7-d4ca-4ca3-9cc3-d5abaeff069f","Type":"ContainerStarted","Data":"ea099a1d74f146c91d713e9334826312b7726ae1dc6d0ce1a041d25dcc71655b"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.058661 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" event={"ID":"f1cbf7ea-2adc-4b59-a731-ee978309766d","Type":"ContainerStarted","Data":"6637caff657a194d714b07c3bbd21452370f8c752203f39fedd124061aa57532"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.058687 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" event={"ID":"f1cbf7ea-2adc-4b59-a731-ee978309766d","Type":"ContainerStarted","Data":"51c73646bbb12ca58709872320d1e91e564743eb304a0408041c6d0df6f9af66"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.060426 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-h2f8n" event={"ID":"f35270c7-ed05-4b4e-bbb4-ad87efb8f940","Type":"ContainerStarted","Data":"51e637f96970fb13ad4cfa4e42aec40294046cbab1fa4f173e9064735ee5454d"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.068282 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" event={"ID":"5eb78f90-4bc7-402f-96d6-ed8bbabb5660","Type":"ContainerStarted","Data":"1da5c7d3ed842b7c1e9800b1d9e86c34dbab400031a092a948c86489bd050fb1"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.068329 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" event={"ID":"5eb78f90-4bc7-402f-96d6-ed8bbabb5660","Type":"ContainerStarted","Data":"c6d6320688f9b3171862182e6883dcfecb4498c1bec76f75755464468903f81d"} Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.070201 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" podStartSLOduration=126.070102455 podStartE2EDuration="2m6.070102455s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.066660414 +0000 UTC m=+145.217127779" watchObservedRunningTime="2026-02-03 09:12:34.070102455 +0000 UTC m=+145.220569830" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.071418 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.071703 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.571694286 +0000 UTC m=+145.722161661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.071859 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r9s2p" podStartSLOduration=126.071851672 podStartE2EDuration="2m6.071851672s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.03872534 +0000 UTC m=+145.189192735" watchObservedRunningTime="2026-02-03 09:12:34.071851672 +0000 UTC m=+145.222319047" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.088609 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sjcqf" podStartSLOduration=125.088465835 podStartE2EDuration="2m5.088465835s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.085567265 +0000 UTC m=+145.236034640" watchObservedRunningTime="2026-02-03 09:12:34.088465835 +0000 UTC m=+145.238933210" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.109748 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-h2f8n" podStartSLOduration=7.109727815 podStartE2EDuration="7.109727815s" podCreationTimestamp="2026-02-03 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.105851428 +0000 UTC m=+145.256318803" watchObservedRunningTime="2026-02-03 09:12:34.109727815 +0000 UTC m=+145.260195200" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.138161 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6q85j" podStartSLOduration=125.138136398 podStartE2EDuration="2m5.138136398s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.136270937 +0000 UTC m=+145.286738322" watchObservedRunningTime="2026-02-03 09:12:34.138136398 +0000 UTC m=+145.288603783" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.156851 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" podStartSLOduration=125.15683242 podStartE2EDuration="2m5.15683242s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.156503127 +0000 UTC m=+145.306970502" watchObservedRunningTime="2026-02-03 09:12:34.15683242 +0000 UTC m=+145.307299795" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.172451 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.173776 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.673742895 +0000 UTC m=+145.824210280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.173958 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.184213 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.684192983 +0000 UTC m=+145.834660359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.188732 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:34 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:34 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:34 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.188787 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.244890 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-25jr4" podStartSLOduration=125.244862335 podStartE2EDuration="2m5.244862335s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.192909755 +0000 UTC m=+145.343377130" watchObservedRunningTime="2026-02-03 09:12:34.244862335 +0000 UTC m=+145.395329730" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.246183 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zgdp2" podStartSLOduration=125.246165434 podStartE2EDuration="2m5.246165434s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.174779194 +0000 UTC m=+145.325246579" watchObservedRunningTime="2026-02-03 09:12:34.246165434 +0000 UTC m=+145.396632819" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.261911 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gsgxt" podStartSLOduration=125.261895284 podStartE2EDuration="2m5.261895284s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.237038017 +0000 UTC m=+145.387505392" watchObservedRunningTime="2026-02-03 09:12:34.261895284 +0000 UTC m=+145.412362659" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.272308 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kwhqd" podStartSLOduration=126.27227838 podStartE2EDuration="2m6.27227838s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.2688682 +0000 UTC m=+145.419335575" watchObservedRunningTime="2026-02-03 09:12:34.27227838 +0000 UTC m=+145.422745755" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.277186 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.277592 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.777577972 +0000 UTC m=+145.928045347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.309906 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" podStartSLOduration=125.309886654 podStartE2EDuration="2m5.309886654s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:34.307640737 +0000 UTC m=+145.458108112" watchObservedRunningTime="2026-02-03 09:12:34.309886654 +0000 UTC m=+145.460354029" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.378648 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.378951 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.878939525 +0000 UTC m=+146.029406900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.479504 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.480109 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:34.98009351 +0000 UTC m=+146.130560885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.580787 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.581281 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.081269786 +0000 UTC m=+146.231737161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.682497 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.682708 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.18267658 +0000 UTC m=+146.333143955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.682774 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.683104 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.183090425 +0000 UTC m=+146.333557800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.735130 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.735305 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.744914 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.784436 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.784598 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.284570793 +0000 UTC m=+146.435038168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.784670 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.785012 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.285000639 +0000 UTC m=+146.435468014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.886225 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.886354 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.386335141 +0000 UTC m=+146.536802526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.886925 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.887241 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.387230185 +0000 UTC m=+146.537697570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:34 crc kubenswrapper[4756]: I0203 09:12:34.987557 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:34 crc kubenswrapper[4756]: E0203 09:12:34.988011 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.487996515 +0000 UTC m=+146.638463890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.074422 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" event={"ID":"c818aba6-e4ed-442e-97c1-b73f5c99e404","Type":"ContainerStarted","Data":"ae984fddd3ecc163d009864ca4c51a5fa4d6d2353e95aae174664351d6b545e1"} Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.077190 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" event={"ID":"4af6f100-b621-407d-9096-1c10d304db9c","Type":"ContainerStarted","Data":"7fe0538204b64932652db2bec353a29bb7df85c03195b821daaa3d5c227e2fde"} Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.080003 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" event={"ID":"1820186d-e8a3-4f09-a6db-01b7bbaa583c","Type":"ContainerStarted","Data":"b06fd5a81315da860ae8680016528829b7b7601d6ceb32d63c766a90cf8a52f8"} Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.080032 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" event={"ID":"1820186d-e8a3-4f09-a6db-01b7bbaa583c","Type":"ContainerStarted","Data":"d0b4d037d96e04c08e82bd80027f134e6e97f67e3c52ba4da19f77f06e101753"} Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.083058 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" event={"ID":"35e10745-2b1b-4843-a815-98bb6f85ca27","Type":"ContainerStarted","Data":"ca30acf45fb9e34c41a38c793c02c05086ccf4b48a75e2f5fd7564f92c76ee83"} Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.084794 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" event={"ID":"6e87ebde-e93e-4cbe-a868-8f09775fd15c","Type":"ContainerStarted","Data":"9e4749036b0d8f78f6ed828b87bf47afb38eb857a2d8504f84f213221f982620"} Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.085134 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.086849 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ctwl6" event={"ID":"3f642171-10cc-4c2d-b596-52b418d20ee8","Type":"ContainerStarted","Data":"287869ac6ee6e1673a0b4729decbf1955588117147793a2cee478aa6d0d1207e"} Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.086874 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.087474 4756 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xn68m container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.087503 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" podUID="3f848653-564d-4a41-ab3b-3e509a42076e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.089063 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.089843 4756 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bltft container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.089990 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" podUID="8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.090143 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.589584417 +0000 UTC m=+146.740051792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.099169 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dtkjs" podStartSLOduration=126.099152832 podStartE2EDuration="2m6.099152832s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:35.09755269 +0000 UTC m=+146.248020065" watchObservedRunningTime="2026-02-03 09:12:35.099152832 +0000 UTC m=+146.249620207" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.100805 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wgxqm" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.106769 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vdsfv" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.125536 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-8jkfm" podStartSLOduration=127.125521356 podStartE2EDuration="2m7.125521356s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:35.122610045 +0000 UTC m=+146.273077420" watchObservedRunningTime="2026-02-03 09:12:35.125521356 +0000 UTC m=+146.275988731" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.144812 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qz4nl" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.168040 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:35 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:35 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:35 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.168596 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.182436 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-jwhwp" podStartSLOduration=126.182421985 podStartE2EDuration="2m6.182421985s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:35.146200374 +0000 UTC m=+146.296667759" watchObservedRunningTime="2026-02-03 09:12:35.182421985 +0000 UTC m=+146.332889360" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.183912 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" podStartSLOduration=126.183904361 podStartE2EDuration="2m6.183904361s" podCreationTimestamp="2026-02-03 09:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:35.180902817 +0000 UTC m=+146.331370192" watchObservedRunningTime="2026-02-03 09:12:35.183904361 +0000 UTC m=+146.334371736" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.190621 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.192653 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.692638044 +0000 UTC m=+146.843105419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.267867 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-ctwl6" podStartSLOduration=8.267847621 podStartE2EDuration="8.267847621s" podCreationTimestamp="2026-02-03 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:35.220685563 +0000 UTC m=+146.371152938" watchObservedRunningTime="2026-02-03 09:12:35.267847621 +0000 UTC m=+146.418314996" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.268087 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" podStartSLOduration=127.268084169 podStartE2EDuration="2m7.268084169s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:35.26414766 +0000 UTC m=+146.414615035" watchObservedRunningTime="2026-02-03 09:12:35.268084169 +0000 UTC m=+146.418551534" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.275316 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.275368 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.276996 4756 patch_prober.go:28] interesting pod/apiserver-76f77b778f-9ghwh container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.277040 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" podUID="1820186d-e8a3-4f09-a6db-01b7bbaa583c" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.293211 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.293512 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.793501918 +0000 UTC m=+146.943969293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.397062 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.397168 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.897151628 +0000 UTC m=+147.047619003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.397417 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.397698 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.897691108 +0000 UTC m=+147.048158483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.498396 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.498636 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.998596074 +0000 UTC m=+147.149063449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.498830 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.499222 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:35.999213778 +0000 UTC m=+147.149681153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.599966 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.600187 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.100153424 +0000 UTC m=+147.250620819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.600480 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.600894 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.100883502 +0000 UTC m=+147.251350957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.701757 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.701947 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.201914762 +0000 UTC m=+147.352382147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.803060 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.803536 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.303513634 +0000 UTC m=+147.453981089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:35 crc kubenswrapper[4756]: I0203 09:12:35.904136 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:35 crc kubenswrapper[4756]: E0203 09:12:35.904451 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.4044245 +0000 UTC m=+147.554891865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.006032 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.006519 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.50650011 +0000 UTC m=+147.656967555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.089498 4756 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8gmxs container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.089574 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" podUID="3e5e0d3e-9130-4986-9369-71439f93c3ac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.090513 4756 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8jhbb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.40:6443/healthz\": context deadline exceeded" start-of-body= Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.090584 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" podUID="1daff10d-ff13-4c5a-9433-d33ade5b6486" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.40:6443/healthz\": context deadline exceeded" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.093765 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sprfw" event={"ID":"6f020886-dc7a-49db-ad89-976ec3f8d771","Type":"ContainerStarted","Data":"b27f0d1d0b6d9d94250bc61fe217a5a9e4da83720da8ee8fc60ac9c1d21ed942"} Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.106863 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.107048 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.607019981 +0000 UTC m=+147.757487356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.107079 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.107091 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.107367 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.607358914 +0000 UTC m=+147.757826319 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.169602 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:36 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:36 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:36 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.169666 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.208114 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.208621 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.708586931 +0000 UTC m=+147.859054306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.209177 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.215914 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.71589505 +0000 UTC m=+147.866362515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.318501 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.318841 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.818824232 +0000 UTC m=+147.969291607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.419645 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.419688 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.419744 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.419784 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.419803 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.420608 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:36.9205809 +0000 UTC m=+148.071048275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.423475 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.427146 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.429965 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.433888 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.434518 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-2mtqs" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.520657 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.521101 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:37.02108316 +0000 UTC m=+148.171550545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.527531 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.531306 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v6bml"] Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.532546 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.534692 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.540552 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.543009 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.577174 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6bml"] Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.622760 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.623062 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:37.123051576 +0000 UTC m=+148.273518951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.704098 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xw49c"] Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.705261 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.719650 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xw49c"] Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.720258 4756 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.720677 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.723897 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.724048 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-utilities\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.724160 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-catalog-content\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.724205 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65dpn\" (UniqueName: \"kubernetes.io/projected/775639e3-1ecc-4ccd-8207-fb6a557e49b3-kube-api-access-65dpn\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.724313 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:37.224295214 +0000 UTC m=+148.374762589 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.783062 4756 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-03T09:12:36.720339184Z","Handler":null,"Name":""} Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.825576 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxcht\" (UniqueName: \"kubernetes.io/projected/95a68714-9c47-4457-a37e-90a084d1a2b0-kube-api-access-qxcht\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.825613 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-utilities\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.825642 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.825660 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-catalog-content\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.825687 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65dpn\" (UniqueName: \"kubernetes.io/projected/775639e3-1ecc-4ccd-8207-fb6a557e49b3-kube-api-access-65dpn\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.825704 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-catalog-content\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.825725 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-utilities\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.826338 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-utilities\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.826605 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:37.326593773 +0000 UTC m=+148.477061148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.826798 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-catalog-content\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.883684 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65dpn\" (UniqueName: \"kubernetes.io/projected/775639e3-1ecc-4ccd-8207-fb6a557e49b3-kube-api-access-65dpn\") pod \"community-operators-v6bml\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.896985 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s2lwz"] Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.898352 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.915611 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s2lwz"] Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.927163 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.927547 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxcht\" (UniqueName: \"kubernetes.io/projected/95a68714-9c47-4457-a37e-90a084d1a2b0-kube-api-access-qxcht\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.927597 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-utilities\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.927769 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-catalog-content\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: E0203 09:12:36.928223 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-03 09:12:37.428204895 +0000 UTC m=+148.578672270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.928341 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-utilities\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.928361 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-catalog-content\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:36 crc kubenswrapper[4756]: I0203 09:12:36.957390 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxcht\" (UniqueName: \"kubernetes.io/projected/95a68714-9c47-4457-a37e-90a084d1a2b0-kube-api-access-qxcht\") pod \"certified-operators-xw49c\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.028974 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-utilities\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.029047 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.029073 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.029102 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-catalog-content\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.029149 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm8kw\" (UniqueName: \"kubernetes.io/projected/a22de99d-c3fa-444b-ae14-7e8de2145e01-kube-api-access-wm8kw\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: E0203 09:12:37.029534 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-03 09:12:37.529520116 +0000 UTC m=+148.679987481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxxsv" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.041975 4756 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.042006 4756 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.058227 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.132634 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.132887 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-catalog-content\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.132940 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm8kw\" (UniqueName: \"kubernetes.io/projected/a22de99d-c3fa-444b-ae14-7e8de2145e01-kube-api-access-wm8kw\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.132996 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-utilities\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.133478 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-utilities\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.133858 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-catalog-content\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.149344 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.151724 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zrjwc"] Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.153301 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.155114 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.171023 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:37 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:37 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:37 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.171074 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.175833 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zrjwc"] Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.185209 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm8kw\" (UniqueName: \"kubernetes.io/projected/a22de99d-c3fa-444b-ae14-7e8de2145e01-kube-api-access-wm8kw\") pod \"community-operators-s2lwz\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.222905 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sprfw" event={"ID":"6f020886-dc7a-49db-ad89-976ec3f8d771","Type":"ContainerStarted","Data":"e0077ad649079fdc2257cef387a0cc7cb6a0a301ec9c80286c301847692d78cf"} Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.222975 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sprfw" event={"ID":"6f020886-dc7a-49db-ad89-976ec3f8d771","Type":"ContainerStarted","Data":"c32de0ff30a791b33293a1d712779e48f54fd1c326f9b39ba66485407bfae1e9"} Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.237293 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.270398 4756 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.270639 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.280771 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-sprfw" podStartSLOduration=10.280756141 podStartE2EDuration="10.280756141s" podCreationTimestamp="2026-02-03 09:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:37.271150195 +0000 UTC m=+148.421617570" watchObservedRunningTime="2026-02-03 09:12:37.280756141 +0000 UTC m=+148.431223516" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.281665 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.282262 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.303670 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.303724 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.304393 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.341299 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-catalog-content\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.341677 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7pd9\" (UniqueName: \"kubernetes.io/projected/a933c882-e11d-46a8-8f67-f77ebc97367c-kube-api-access-h7pd9\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.341766 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-utilities\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.345676 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.362294 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxxsv\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.442784 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.442864 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.442895 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-catalog-content\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.442923 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7pd9\" (UniqueName: \"kubernetes.io/projected/a933c882-e11d-46a8-8f67-f77ebc97367c-kube-api-access-h7pd9\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.442951 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-utilities\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.444614 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-utilities\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.444926 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-catalog-content\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.497289 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7pd9\" (UniqueName: \"kubernetes.io/projected/a933c882-e11d-46a8-8f67-f77ebc97367c-kube-api-access-h7pd9\") pod \"certified-operators-zrjwc\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.506821 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.545805 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.546210 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.546631 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.589895 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.655865 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.665688 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.667648 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:37 crc kubenswrapper[4756]: W0203 09:12:37.735924 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-a0e6bf9fa58917c1658d24371bddf001707e22e39f47d1357e5a65f1c7a8482e WatchSource:0}: Error finding container a0e6bf9fa58917c1658d24371bddf001707e22e39f47d1357e5a65f1c7a8482e: Status 404 returned error can't find the container with id a0e6bf9fa58917c1658d24371bddf001707e22e39f47d1357e5a65f1c7a8482e Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.738522 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xw49c"] Feb 03 09:12:37 crc kubenswrapper[4756]: I0203 09:12:37.865752 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6bml"] Feb 03 09:12:37 crc kubenswrapper[4756]: W0203 09:12:37.876417 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod775639e3_1ecc_4ccd_8207_fb6a557e49b3.slice/crio-072c6adc1547fdbf15aa241e57909345bec616cfa902b47d60cf7655380defa0 WatchSource:0}: Error finding container 072c6adc1547fdbf15aa241e57909345bec616cfa902b47d60cf7655380defa0: Status 404 returned error can't find the container with id 072c6adc1547fdbf15aa241e57909345bec616cfa902b47d60cf7655380defa0 Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.112898 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s2lwz"] Feb 03 09:12:38 crc kubenswrapper[4756]: W0203 09:12:38.139951 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda22de99d_c3fa_444b_ae14_7e8de2145e01.slice/crio-877fb99d49a261ca592ecd832701086142134dc2eb632deb97aae3396c81a42f WatchSource:0}: Error finding container 877fb99d49a261ca592ecd832701086142134dc2eb632deb97aae3396c81a42f: Status 404 returned error can't find the container with id 877fb99d49a261ca592ecd832701086142134dc2eb632deb97aae3396c81a42f Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.174721 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:38 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:38 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:38 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.174781 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.212825 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zrjwc"] Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.283193 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6bml" event={"ID":"775639e3-1ecc-4ccd-8207-fb6a557e49b3","Type":"ContainerStarted","Data":"072c6adc1547fdbf15aa241e57909345bec616cfa902b47d60cf7655380defa0"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.309204 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw49c" event={"ID":"95a68714-9c47-4457-a37e-90a084d1a2b0","Type":"ContainerStarted","Data":"6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.309282 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw49c" event={"ID":"95a68714-9c47-4457-a37e-90a084d1a2b0","Type":"ContainerStarted","Data":"6ddab397bc19c427cbf60f38f2590c655a17376aaab584a70c17c5f2327321bf"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.313970 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.314956 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.351025 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3924571ee5001f8a0ec5791d4f1faef6625c4851e1f0deed1ea0ba9718d28931"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.351078 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1e9336fe4c8e63140f35d5855438bb0667b772863e3c10887a9370ff79b5874b"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.351751 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.356095 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxxsv"] Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.409945 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"af964925ce322473cf765826b1ff8d8fc20215abf986808d6c033c3b94055442"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.410000 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"93f16ad8ba1d4f4fc510d0d163e2c6b9b840e95d3e8d63c69b7280339cfe8b0d"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.414717 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2lwz" event={"ID":"a22de99d-c3fa-444b-ae14-7e8de2145e01","Type":"ContainerStarted","Data":"877fb99d49a261ca592ecd832701086142134dc2eb632deb97aae3396c81a42f"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.417038 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f62e2acadc40cd332cd3347f6e332e4a3b80627e9a6365a86f3ed56228e493ae"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.417072 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a0e6bf9fa58917c1658d24371bddf001707e22e39f47d1357e5a65f1c7a8482e"} Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.495734 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.532169 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s64pq"] Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.533676 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.535878 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s64pq"] Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.548043 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.680092 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7825\" (UniqueName: \"kubernetes.io/projected/fc312654-25ec-4cf1-b35b-ea49aa157603-kube-api-access-f7825\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.680379 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-catalog-content\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.680414 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-utilities\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.783005 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7825\" (UniqueName: \"kubernetes.io/projected/fc312654-25ec-4cf1-b35b-ea49aa157603-kube-api-access-f7825\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.783071 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-catalog-content\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.783109 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-utilities\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.783722 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-utilities\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.784099 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-catalog-content\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.812462 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7825\" (UniqueName: \"kubernetes.io/projected/fc312654-25ec-4cf1-b35b-ea49aa157603-kube-api-access-f7825\") pod \"redhat-marketplace-s64pq\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.900807 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-frm8n"] Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.902090 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.911374 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:12:38 crc kubenswrapper[4756]: I0203 09:12:38.915425 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-frm8n"] Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.087250 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-catalog-content\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.087630 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndnt8\" (UniqueName: \"kubernetes.io/projected/cdb0e910-6e83-440e-bfb6-5e78e960c99d-kube-api-access-ndnt8\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.087656 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-utilities\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.158411 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s64pq"] Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.162857 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:39 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:39 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:39 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.162925 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:39 crc kubenswrapper[4756]: W0203 09:12:39.169926 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc312654_25ec_4cf1_b35b_ea49aa157603.slice/crio-c4301705640587639a286214d9eead21b0ab5aaf2c99f2705a0da284980fd69c WatchSource:0}: Error finding container c4301705640587639a286214d9eead21b0ab5aaf2c99f2705a0da284980fd69c: Status 404 returned error can't find the container with id c4301705640587639a286214d9eead21b0ab5aaf2c99f2705a0da284980fd69c Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.189310 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-catalog-content\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.189420 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndnt8\" (UniqueName: \"kubernetes.io/projected/cdb0e910-6e83-440e-bfb6-5e78e960c99d-kube-api-access-ndnt8\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.189479 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-utilities\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.189929 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-utilities\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.192916 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-catalog-content\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.206937 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndnt8\" (UniqueName: \"kubernetes.io/projected/cdb0e910-6e83-440e-bfb6-5e78e960c99d-kube-api-access-ndnt8\") pod \"redhat-marketplace-frm8n\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.268038 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.440067 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" event={"ID":"e579b706-5cc4-4a92-94f2-b22ad2135ffc","Type":"ContainerStarted","Data":"3fa1da28099b8974f60dbf71def0ff3fd5c772f1ef99b6af2a6ad2dfd126e834"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.440124 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" event={"ID":"e579b706-5cc4-4a92-94f2-b22ad2135ffc","Type":"ContainerStarted","Data":"fe46f7277a65b7646abd62fc32a5c9d6ea9a3c924f57dbd7346feef9304a4771"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.440711 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.442668 4756 generic.go:334] "Generic (PLEG): container finished" podID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerID="6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433" exitCode=0 Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.442735 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw49c" event={"ID":"95a68714-9c47-4457-a37e-90a084d1a2b0","Type":"ContainerDied","Data":"6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.445144 4756 generic.go:334] "Generic (PLEG): container finished" podID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerID="c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409" exitCode=0 Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.445203 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6bml" event={"ID":"775639e3-1ecc-4ccd-8207-fb6a557e49b3","Type":"ContainerDied","Data":"c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.447212 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-frm8n"] Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.447510 4756 generic.go:334] "Generic (PLEG): container finished" podID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerID="481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2" exitCode=0 Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.447581 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrjwc" event={"ID":"a933c882-e11d-46a8-8f67-f77ebc97367c","Type":"ContainerDied","Data":"481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.447603 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrjwc" event={"ID":"a933c882-e11d-46a8-8f67-f77ebc97367c","Type":"ContainerStarted","Data":"704345cc223a4f28260edcdcf3d772902276d126d441a7f5e81fddec3c22e9ae"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.451064 4756 generic.go:334] "Generic (PLEG): container finished" podID="e742c048-9564-4b75-986b-a34639494020" containerID="ff07b217cc643322909bb9ab3d7a77b96673d00b9ba816c5d024d1b3056c44a2" exitCode=0 Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.451095 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" event={"ID":"e742c048-9564-4b75-986b-a34639494020","Type":"ContainerDied","Data":"ff07b217cc643322909bb9ab3d7a77b96673d00b9ba816c5d024d1b3056c44a2"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.457923 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"057d87cb-bf18-4ad8-8b53-ddb060fb72d8","Type":"ContainerStarted","Data":"2d098a49c36fd3e445818ecb3ca549b93908e647b7c792fb0be5a5011b4f9553"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.457978 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"057d87cb-bf18-4ad8-8b53-ddb060fb72d8","Type":"ContainerStarted","Data":"9dadd691b372438eaf4a907d90b3bea73864343e7eac95dfb3247de79a29e7b1"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.460656 4756 generic.go:334] "Generic (PLEG): container finished" podID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerID="ec1737f649380811b51c94caa9ea9e6b1c07426d6c0e12da12284e8faa6443e8" exitCode=0 Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.460728 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2lwz" event={"ID":"a22de99d-c3fa-444b-ae14-7e8de2145e01","Type":"ContainerDied","Data":"ec1737f649380811b51c94caa9ea9e6b1c07426d6c0e12da12284e8faa6443e8"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.468568 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.469247 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.470249 4756 generic.go:334] "Generic (PLEG): container finished" podID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerID="76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb" exitCode=0 Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.471013 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s64pq" event={"ID":"fc312654-25ec-4cf1-b35b-ea49aa157603","Type":"ContainerDied","Data":"76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.471045 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s64pq" event={"ID":"fc312654-25ec-4cf1-b35b-ea49aa157603","Type":"ContainerStarted","Data":"c4301705640587639a286214d9eead21b0ab5aaf2c99f2705a0da284980fd69c"} Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.475165 4756 patch_prober.go:28] interesting pod/console-f9d7485db-6qrwb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.34:8443/health\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.475202 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6qrwb" podUID="4dad28d0-454d-434b-b111-137c215aea4a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.34:8443/health\": dial tcp 10.217.0.34:8443: connect: connection refused" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.488815 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" podStartSLOduration=131.488795907 podStartE2EDuration="2m11.488795907s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:39.468823946 +0000 UTC m=+150.619291341" watchObservedRunningTime="2026-02-03 09:12:39.488795907 +0000 UTC m=+150.639263282" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.530929 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.530891342 podStartE2EDuration="2.530891342s" podCreationTimestamp="2026-02-03 09:12:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:39.529040361 +0000 UTC m=+150.679507756" watchObservedRunningTime="2026-02-03 09:12:39.530891342 +0000 UTC m=+150.681358717" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.549062 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-nzcq8" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.883572 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pz2qm"] Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.884712 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.887155 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 03 09:12:39 crc kubenswrapper[4756]: I0203 09:12:39.899942 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pz2qm"] Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.012271 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg9ls\" (UniqueName: \"kubernetes.io/projected/78ac0913-94cb-4f26-934d-bd005399d428-kube-api-access-sg9ls\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.012416 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-catalog-content\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.012485 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-utilities\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.113777 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-catalog-content\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.113859 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-utilities\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.113921 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg9ls\" (UniqueName: \"kubernetes.io/projected/78ac0913-94cb-4f26-934d-bd005399d428-kube-api-access-sg9ls\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.115097 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-utilities\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.115354 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-catalog-content\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.118320 4756 patch_prober.go:28] interesting pod/downloads-7954f5f757-7nst7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.119650 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7nst7" podUID="142cdcb7-ccff-4a63-99c5-e857b8eaa4df" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.120573 4756 patch_prober.go:28] interesting pod/downloads-7954f5f757-7nst7 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.120607 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7nst7" podUID="142cdcb7-ccff-4a63-99c5-e857b8eaa4df" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.137475 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg9ls\" (UniqueName: \"kubernetes.io/projected/78ac0913-94cb-4f26-934d-bd005399d428-kube-api-access-sg9ls\") pod \"redhat-operators-pz2qm\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.160720 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.167137 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:40 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:40 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:40 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.167262 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.217518 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.285986 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qpmvh"] Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.287223 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.303803 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.308008 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qpmvh"] Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.311327 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-9ghwh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.420871 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trd5n\" (UniqueName: \"kubernetes.io/projected/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-kube-api-access-trd5n\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.420953 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-utilities\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.421036 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-catalog-content\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.483221 4756 generic.go:334] "Generic (PLEG): container finished" podID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerID="93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd" exitCode=0 Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.483302 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frm8n" event={"ID":"cdb0e910-6e83-440e-bfb6-5e78e960c99d","Type":"ContainerDied","Data":"93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd"} Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.483339 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frm8n" event={"ID":"cdb0e910-6e83-440e-bfb6-5e78e960c99d","Type":"ContainerStarted","Data":"2aabdeccb55adc8a6cf051e250d633cfd6de3c59b15e36a0f7be641358b0e478"} Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.486021 4756 generic.go:334] "Generic (PLEG): container finished" podID="057d87cb-bf18-4ad8-8b53-ddb060fb72d8" containerID="2d098a49c36fd3e445818ecb3ca549b93908e647b7c792fb0be5a5011b4f9553" exitCode=0 Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.486065 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"057d87cb-bf18-4ad8-8b53-ddb060fb72d8","Type":"ContainerDied","Data":"2d098a49c36fd3e445818ecb3ca549b93908e647b7c792fb0be5a5011b4f9553"} Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.522289 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trd5n\" (UniqueName: \"kubernetes.io/projected/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-kube-api-access-trd5n\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.522363 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-utilities\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.522420 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-catalog-content\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.523065 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-catalog-content\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.523667 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-utilities\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.561327 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trd5n\" (UniqueName: \"kubernetes.io/projected/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-kube-api-access-trd5n\") pod \"redhat-operators-qpmvh\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.578593 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.620666 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8gmxs" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.641949 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:12:40 crc kubenswrapper[4756]: I0203 09:12:40.776579 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pz2qm"] Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.062049 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qpmvh"] Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.064091 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.151429 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e742c048-9564-4b75-986b-a34639494020-secret-volume\") pod \"e742c048-9564-4b75-986b-a34639494020\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.151501 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9tz6\" (UniqueName: \"kubernetes.io/projected/e742c048-9564-4b75-986b-a34639494020-kube-api-access-s9tz6\") pod \"e742c048-9564-4b75-986b-a34639494020\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.151528 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e742c048-9564-4b75-986b-a34639494020-config-volume\") pod \"e742c048-9564-4b75-986b-a34639494020\" (UID: \"e742c048-9564-4b75-986b-a34639494020\") " Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.152778 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e742c048-9564-4b75-986b-a34639494020-config-volume" (OuterVolumeSpecName: "config-volume") pod "e742c048-9564-4b75-986b-a34639494020" (UID: "e742c048-9564-4b75-986b-a34639494020"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.158897 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e742c048-9564-4b75-986b-a34639494020-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e742c048-9564-4b75-986b-a34639494020" (UID: "e742c048-9564-4b75-986b-a34639494020"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.159403 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e742c048-9564-4b75-986b-a34639494020-kube-api-access-s9tz6" (OuterVolumeSpecName: "kube-api-access-s9tz6") pod "e742c048-9564-4b75-986b-a34639494020" (UID: "e742c048-9564-4b75-986b-a34639494020"). InnerVolumeSpecName "kube-api-access-s9tz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.164200 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:41 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:41 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:41 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.164276 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.252717 4756 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e742c048-9564-4b75-986b-a34639494020-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.252747 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9tz6\" (UniqueName: \"kubernetes.io/projected/e742c048-9564-4b75-986b-a34639494020-kube-api-access-s9tz6\") on node \"crc\" DevicePath \"\"" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.252757 4756 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e742c048-9564-4b75-986b-a34639494020-config-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.505642 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.506002 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z" event={"ID":"e742c048-9564-4b75-986b-a34639494020","Type":"ContainerDied","Data":"e8f2ab8f5bb9faa51ec8e7a09f426a6bd186e6a9b166f43f6101f8104e86b746"} Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.506045 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8f2ab8f5bb9faa51ec8e7a09f426a6bd186e6a9b166f43f6101f8104e86b746" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.514743 4756 generic.go:334] "Generic (PLEG): container finished" podID="78ac0913-94cb-4f26-934d-bd005399d428" containerID="bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d" exitCode=0 Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.514849 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz2qm" event={"ID":"78ac0913-94cb-4f26-934d-bd005399d428","Type":"ContainerDied","Data":"bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d"} Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.514885 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz2qm" event={"ID":"78ac0913-94cb-4f26-934d-bd005399d428","Type":"ContainerStarted","Data":"367617e4713272129bdb888c9eed5b28230d7259bf332e9b8e67c0cc16f99e60"} Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.524478 4756 generic.go:334] "Generic (PLEG): container finished" podID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerID="6a92eb4bbe0bd609976e45303572866ca0751213d2da93649e90354eb7e3291e" exitCode=0 Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.524568 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpmvh" event={"ID":"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37","Type":"ContainerDied","Data":"6a92eb4bbe0bd609976e45303572866ca0751213d2da93649e90354eb7e3291e"} Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.524615 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpmvh" event={"ID":"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37","Type":"ContainerStarted","Data":"a5dcd811d66ce87825a7413c5e9f343716a305c64854bc27023683cb1670793e"} Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.807481 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.837002 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 03 09:12:41 crc kubenswrapper[4756]: E0203 09:12:41.837296 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057d87cb-bf18-4ad8-8b53-ddb060fb72d8" containerName="pruner" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.837312 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="057d87cb-bf18-4ad8-8b53-ddb060fb72d8" containerName="pruner" Feb 03 09:12:41 crc kubenswrapper[4756]: E0203 09:12:41.837339 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e742c048-9564-4b75-986b-a34639494020" containerName="collect-profiles" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.837346 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e742c048-9564-4b75-986b-a34639494020" containerName="collect-profiles" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.837437 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="057d87cb-bf18-4ad8-8b53-ddb060fb72d8" containerName="pruner" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.837460 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="e742c048-9564-4b75-986b-a34639494020" containerName="collect-profiles" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.838636 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.841073 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.841981 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.860774 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.967684 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kube-api-access\") pod \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\" (UID: \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\") " Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.967849 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kubelet-dir\") pod \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\" (UID: \"057d87cb-bf18-4ad8-8b53-ddb060fb72d8\") " Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.968067 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80e9f32d-1331-4e2d-bff1-153337d89141-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"80e9f32d-1331-4e2d-bff1-153337d89141\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.968173 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80e9f32d-1331-4e2d-bff1-153337d89141-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"80e9f32d-1331-4e2d-bff1-153337d89141\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.975958 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "057d87cb-bf18-4ad8-8b53-ddb060fb72d8" (UID: "057d87cb-bf18-4ad8-8b53-ddb060fb72d8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:12:41 crc kubenswrapper[4756]: I0203 09:12:41.990828 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "057d87cb-bf18-4ad8-8b53-ddb060fb72d8" (UID: "057d87cb-bf18-4ad8-8b53-ddb060fb72d8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.070366 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80e9f32d-1331-4e2d-bff1-153337d89141-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"80e9f32d-1331-4e2d-bff1-153337d89141\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.070502 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80e9f32d-1331-4e2d-bff1-153337d89141-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"80e9f32d-1331-4e2d-bff1-153337d89141\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.070753 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80e9f32d-1331-4e2d-bff1-153337d89141-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"80e9f32d-1331-4e2d-bff1-153337d89141\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.070882 4756 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.070905 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/057d87cb-bf18-4ad8-8b53-ddb060fb72d8-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.108771 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80e9f32d-1331-4e2d-bff1-153337d89141-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"80e9f32d-1331-4e2d-bff1-153337d89141\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.162467 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:42 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:42 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:42 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.162523 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.175056 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.524665 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.542151 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.542139 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"057d87cb-bf18-4ad8-8b53-ddb060fb72d8","Type":"ContainerDied","Data":"9dadd691b372438eaf4a907d90b3bea73864343e7eac95dfb3247de79a29e7b1"} Feb 03 09:12:42 crc kubenswrapper[4756]: I0203 09:12:42.542289 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dadd691b372438eaf4a907d90b3bea73864343e7eac95dfb3247de79a29e7b1" Feb 03 09:12:42 crc kubenswrapper[4756]: W0203 09:12:42.552171 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod80e9f32d_1331_4e2d_bff1_153337d89141.slice/crio-e72cddbb2c3dc42aa97a6f806c281b30de3cbed0f13f97f701f6d8c2e9243488 WatchSource:0}: Error finding container e72cddbb2c3dc42aa97a6f806c281b30de3cbed0f13f97f701f6d8c2e9243488: Status 404 returned error can't find the container with id e72cddbb2c3dc42aa97a6f806c281b30de3cbed0f13f97f701f6d8c2e9243488 Feb 03 09:12:43 crc kubenswrapper[4756]: I0203 09:12:43.163208 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:43 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:43 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:43 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:43 crc kubenswrapper[4756]: I0203 09:12:43.163456 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:43 crc kubenswrapper[4756]: I0203 09:12:43.562478 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"80e9f32d-1331-4e2d-bff1-153337d89141","Type":"ContainerStarted","Data":"b7d3ab75ad7b36fc51bdd1d90d45a20a040b4498de73f38967be207cc961f00f"} Feb 03 09:12:43 crc kubenswrapper[4756]: I0203 09:12:43.562544 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"80e9f32d-1331-4e2d-bff1-153337d89141","Type":"ContainerStarted","Data":"e72cddbb2c3dc42aa97a6f806c281b30de3cbed0f13f97f701f6d8c2e9243488"} Feb 03 09:12:43 crc kubenswrapper[4756]: I0203 09:12:43.566089 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:12:43 crc kubenswrapper[4756]: I0203 09:12:43.566142 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:12:43 crc kubenswrapper[4756]: I0203 09:12:43.587306 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.5872826079999998 podStartE2EDuration="2.587282608s" podCreationTimestamp="2026-02-03 09:12:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:12:43.573607126 +0000 UTC m=+154.724074511" watchObservedRunningTime="2026-02-03 09:12:43.587282608 +0000 UTC m=+154.737749983" Feb 03 09:12:44 crc kubenswrapper[4756]: I0203 09:12:44.163869 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:44 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:44 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:44 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:44 crc kubenswrapper[4756]: I0203 09:12:44.163938 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:44 crc kubenswrapper[4756]: I0203 09:12:44.613336 4756 generic.go:334] "Generic (PLEG): container finished" podID="80e9f32d-1331-4e2d-bff1-153337d89141" containerID="b7d3ab75ad7b36fc51bdd1d90d45a20a040b4498de73f38967be207cc961f00f" exitCode=0 Feb 03 09:12:44 crc kubenswrapper[4756]: I0203 09:12:44.613410 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"80e9f32d-1331-4e2d-bff1-153337d89141","Type":"ContainerDied","Data":"b7d3ab75ad7b36fc51bdd1d90d45a20a040b4498de73f38967be207cc961f00f"} Feb 03 09:12:45 crc kubenswrapper[4756]: I0203 09:12:45.163700 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:45 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:45 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:45 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:45 crc kubenswrapper[4756]: I0203 09:12:45.164193 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:45 crc kubenswrapper[4756]: I0203 09:12:45.649638 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-ctwl6" Feb 03 09:12:46 crc kubenswrapper[4756]: I0203 09:12:46.163141 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:46 crc kubenswrapper[4756]: [-]has-synced failed: reason withheld Feb 03 09:12:46 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:46 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:46 crc kubenswrapper[4756]: I0203 09:12:46.163227 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:47 crc kubenswrapper[4756]: I0203 09:12:47.164465 4756 patch_prober.go:28] interesting pod/router-default-5444994796-66w8b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 03 09:12:47 crc kubenswrapper[4756]: [+]has-synced ok Feb 03 09:12:47 crc kubenswrapper[4756]: [+]process-running ok Feb 03 09:12:47 crc kubenswrapper[4756]: healthz check failed Feb 03 09:12:47 crc kubenswrapper[4756]: I0203 09:12:47.164729 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-66w8b" podUID="625ab6f6-0677-42a9-871d-5f935d03a9b7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 03 09:12:48 crc kubenswrapper[4756]: I0203 09:12:48.163622 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:48 crc kubenswrapper[4756]: I0203 09:12:48.166903 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-66w8b" Feb 03 09:12:49 crc kubenswrapper[4756]: I0203 09:12:49.468015 4756 patch_prober.go:28] interesting pod/console-f9d7485db-6qrwb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.34:8443/health\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Feb 03 09:12:49 crc kubenswrapper[4756]: I0203 09:12:49.468393 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6qrwb" podUID="4dad28d0-454d-434b-b111-137c215aea4a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.34:8443/health\": dial tcp 10.217.0.34:8443: connect: connection refused" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.129594 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-7nst7" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.598283 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.669609 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"80e9f32d-1331-4e2d-bff1-153337d89141","Type":"ContainerDied","Data":"e72cddbb2c3dc42aa97a6f806c281b30de3cbed0f13f97f701f6d8c2e9243488"} Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.669874 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e72cddbb2c3dc42aa97a6f806c281b30de3cbed0f13f97f701f6d8c2e9243488" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.669926 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.718172 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80e9f32d-1331-4e2d-bff1-153337d89141-kube-api-access\") pod \"80e9f32d-1331-4e2d-bff1-153337d89141\" (UID: \"80e9f32d-1331-4e2d-bff1-153337d89141\") " Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.718278 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80e9f32d-1331-4e2d-bff1-153337d89141-kubelet-dir\") pod \"80e9f32d-1331-4e2d-bff1-153337d89141\" (UID: \"80e9f32d-1331-4e2d-bff1-153337d89141\") " Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.718340 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/80e9f32d-1331-4e2d-bff1-153337d89141-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "80e9f32d-1331-4e2d-bff1-153337d89141" (UID: "80e9f32d-1331-4e2d-bff1-153337d89141"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.718862 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.719026 4756 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/80e9f32d-1331-4e2d-bff1-153337d89141-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.723734 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e9f32d-1331-4e2d-bff1-153337d89141-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "80e9f32d-1331-4e2d-bff1-153337d89141" (UID: "80e9f32d-1331-4e2d-bff1-153337d89141"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.723811 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/717b7410-dd64-44cb-ba9b-3436d82ebb95-metrics-certs\") pod \"network-metrics-daemon-k6pzt\" (UID: \"717b7410-dd64-44cb-ba9b-3436d82ebb95\") " pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.819704 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/80e9f32d-1331-4e2d-bff1-153337d89141-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 03 09:12:50 crc kubenswrapper[4756]: I0203 09:12:50.949306 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k6pzt" Feb 03 09:12:52 crc kubenswrapper[4756]: I0203 09:12:52.279080 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xn68m"] Feb 03 09:12:52 crc kubenswrapper[4756]: I0203 09:12:52.279618 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" podUID="3f848653-564d-4a41-ab3b-3e509a42076e" containerName="controller-manager" containerID="cri-o://86c4adbb56c70b5e3b7cbab7f5522d2218131fe50218672931f616f916548e18" gracePeriod=30 Feb 03 09:12:52 crc kubenswrapper[4756]: I0203 09:12:52.300252 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q"] Feb 03 09:12:52 crc kubenswrapper[4756]: I0203 09:12:52.300524 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" podUID="d61f6dad-2b6f-475c-8516-127db0c852c2" containerName="route-controller-manager" containerID="cri-o://5a013ce0da93436a9256ffa45255470f734a9b1e895a504aec1decdd199d08a5" gracePeriod=30 Feb 03 09:12:52 crc kubenswrapper[4756]: I0203 09:12:52.682555 4756 generic.go:334] "Generic (PLEG): container finished" podID="d61f6dad-2b6f-475c-8516-127db0c852c2" containerID="5a013ce0da93436a9256ffa45255470f734a9b1e895a504aec1decdd199d08a5" exitCode=0 Feb 03 09:12:52 crc kubenswrapper[4756]: I0203 09:12:52.682616 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" event={"ID":"d61f6dad-2b6f-475c-8516-127db0c852c2","Type":"ContainerDied","Data":"5a013ce0da93436a9256ffa45255470f734a9b1e895a504aec1decdd199d08a5"} Feb 03 09:12:52 crc kubenswrapper[4756]: I0203 09:12:52.683963 4756 generic.go:334] "Generic (PLEG): container finished" podID="3f848653-564d-4a41-ab3b-3e509a42076e" containerID="86c4adbb56c70b5e3b7cbab7f5522d2218131fe50218672931f616f916548e18" exitCode=0 Feb 03 09:12:52 crc kubenswrapper[4756]: I0203 09:12:52.683991 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" event={"ID":"3f848653-564d-4a41-ab3b-3e509a42076e","Type":"ContainerDied","Data":"86c4adbb56c70b5e3b7cbab7f5522d2218131fe50218672931f616f916548e18"} Feb 03 09:12:57 crc kubenswrapper[4756]: I0203 09:12:57.661664 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:12:59 crc kubenswrapper[4756]: I0203 09:12:59.895825 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:12:59 crc kubenswrapper[4756]: I0203 09:12:59.901313 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:13:01 crc kubenswrapper[4756]: I0203 09:13:01.236671 4756 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-hzt8q container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 03 09:13:01 crc kubenswrapper[4756]: I0203 09:13:01.237052 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" podUID="d61f6dad-2b6f-475c-8516-127db0c852c2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 03 09:13:01 crc kubenswrapper[4756]: I0203 09:13:01.589617 4756 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xn68m container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": context deadline exceeded" start-of-body= Feb 03 09:13:01 crc kubenswrapper[4756]: I0203 09:13:01.589726 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" podUID="3f848653-564d-4a41-ab3b-3e509a42076e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": context deadline exceeded" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.764587 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" event={"ID":"d61f6dad-2b6f-475c-8516-127db0c852c2","Type":"ContainerDied","Data":"73348d585e9315b5978e593697163d9751f2a34edccf501897f236c347d7b6ca"} Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.765115 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73348d585e9315b5978e593697163d9751f2a34edccf501897f236c347d7b6ca" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.766015 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" event={"ID":"3f848653-564d-4a41-ab3b-3e509a42076e","Type":"ContainerDied","Data":"5470dcebe391f9bbfbada4d47fc184f66ee17e2875571918a072348d6b0ed4e8"} Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.766055 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5470dcebe391f9bbfbada4d47fc184f66ee17e2875571918a072348d6b0ed4e8" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.811035 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.822577 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.861093 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9"] Feb 03 09:13:06 crc kubenswrapper[4756]: E0203 09:13:06.862172 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e9f32d-1331-4e2d-bff1-153337d89141" containerName="pruner" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.862273 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e9f32d-1331-4e2d-bff1-153337d89141" containerName="pruner" Feb 03 09:13:06 crc kubenswrapper[4756]: E0203 09:13:06.862375 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f848653-564d-4a41-ab3b-3e509a42076e" containerName="controller-manager" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.862701 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f848653-564d-4a41-ab3b-3e509a42076e" containerName="controller-manager" Feb 03 09:13:06 crc kubenswrapper[4756]: E0203 09:13:06.862865 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61f6dad-2b6f-475c-8516-127db0c852c2" containerName="route-controller-manager" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.863082 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61f6dad-2b6f-475c-8516-127db0c852c2" containerName="route-controller-manager" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.863416 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d61f6dad-2b6f-475c-8516-127db0c852c2" containerName="route-controller-manager" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.864117 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e9f32d-1331-4e2d-bff1-153337d89141" containerName="pruner" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.864216 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f848653-564d-4a41-ab3b-3e509a42076e" containerName="controller-manager" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.865570 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.880273 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9"] Feb 03 09:13:06 crc kubenswrapper[4756]: E0203 09:13:06.884144 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 03 09:13:06 crc kubenswrapper[4756]: E0203 09:13:06.885541 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h7pd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zrjwc_openshift-marketplace(a933c882-e11d-46a8-8f67-f77ebc97367c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 03 09:13:06 crc kubenswrapper[4756]: E0203 09:13:06.887007 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zrjwc" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949211 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-client-ca\") pod \"d61f6dad-2b6f-475c-8516-127db0c852c2\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949241 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-client-ca\") pod \"3f848653-564d-4a41-ab3b-3e509a42076e\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949261 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-config\") pod \"3f848653-564d-4a41-ab3b-3e509a42076e\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949348 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rnf7\" (UniqueName: \"kubernetes.io/projected/d61f6dad-2b6f-475c-8516-127db0c852c2-kube-api-access-2rnf7\") pod \"d61f6dad-2b6f-475c-8516-127db0c852c2\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949367 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d61f6dad-2b6f-475c-8516-127db0c852c2-serving-cert\") pod \"d61f6dad-2b6f-475c-8516-127db0c852c2\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949388 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt5dh\" (UniqueName: \"kubernetes.io/projected/3f848653-564d-4a41-ab3b-3e509a42076e-kube-api-access-jt5dh\") pod \"3f848653-564d-4a41-ab3b-3e509a42076e\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949403 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f848653-564d-4a41-ab3b-3e509a42076e-serving-cert\") pod \"3f848653-564d-4a41-ab3b-3e509a42076e\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949428 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-proxy-ca-bundles\") pod \"3f848653-564d-4a41-ab3b-3e509a42076e\" (UID: \"3f848653-564d-4a41-ab3b-3e509a42076e\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949470 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-config\") pod \"d61f6dad-2b6f-475c-8516-127db0c852c2\" (UID: \"d61f6dad-2b6f-475c-8516-127db0c852c2\") " Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949595 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-proxy-ca-bundles\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949636 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-client-ca\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949664 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-config\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949684 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/637ecd17-93b8-41e7-83f4-210dc564be35-serving-cert\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.949705 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxcqc\" (UniqueName: \"kubernetes.io/projected/637ecd17-93b8-41e7-83f4-210dc564be35-kube-api-access-dxcqc\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.950437 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-client-ca" (OuterVolumeSpecName: "client-ca") pod "d61f6dad-2b6f-475c-8516-127db0c852c2" (UID: "d61f6dad-2b6f-475c-8516-127db0c852c2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.950516 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-client-ca" (OuterVolumeSpecName: "client-ca") pod "3f848653-564d-4a41-ab3b-3e509a42076e" (UID: "3f848653-564d-4a41-ab3b-3e509a42076e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.951130 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-config" (OuterVolumeSpecName: "config") pod "3f848653-564d-4a41-ab3b-3e509a42076e" (UID: "3f848653-564d-4a41-ab3b-3e509a42076e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.951131 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-config" (OuterVolumeSpecName: "config") pod "d61f6dad-2b6f-475c-8516-127db0c852c2" (UID: "d61f6dad-2b6f-475c-8516-127db0c852c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.951586 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3f848653-564d-4a41-ab3b-3e509a42076e" (UID: "3f848653-564d-4a41-ab3b-3e509a42076e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.962243 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f848653-564d-4a41-ab3b-3e509a42076e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3f848653-564d-4a41-ab3b-3e509a42076e" (UID: "3f848653-564d-4a41-ab3b-3e509a42076e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.964435 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d61f6dad-2b6f-475c-8516-127db0c852c2-kube-api-access-2rnf7" (OuterVolumeSpecName: "kube-api-access-2rnf7") pod "d61f6dad-2b6f-475c-8516-127db0c852c2" (UID: "d61f6dad-2b6f-475c-8516-127db0c852c2"). InnerVolumeSpecName "kube-api-access-2rnf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.968654 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f848653-564d-4a41-ab3b-3e509a42076e-kube-api-access-jt5dh" (OuterVolumeSpecName: "kube-api-access-jt5dh") pod "3f848653-564d-4a41-ab3b-3e509a42076e" (UID: "3f848653-564d-4a41-ab3b-3e509a42076e"). InnerVolumeSpecName "kube-api-access-jt5dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:06 crc kubenswrapper[4756]: I0203 09:13:06.968861 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d61f6dad-2b6f-475c-8516-127db0c852c2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d61f6dad-2b6f-475c-8516-127db0c852c2" (UID: "d61f6dad-2b6f-475c-8516-127db0c852c2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.050948 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-proxy-ca-bundles\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051024 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-client-ca\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051064 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-config\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051091 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/637ecd17-93b8-41e7-83f4-210dc564be35-serving-cert\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051124 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxcqc\" (UniqueName: \"kubernetes.io/projected/637ecd17-93b8-41e7-83f4-210dc564be35-kube-api-access-dxcqc\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051215 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rnf7\" (UniqueName: \"kubernetes.io/projected/d61f6dad-2b6f-475c-8516-127db0c852c2-kube-api-access-2rnf7\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051254 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d61f6dad-2b6f-475c-8516-127db0c852c2-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051266 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt5dh\" (UniqueName: \"kubernetes.io/projected/3f848653-564d-4a41-ab3b-3e509a42076e-kube-api-access-jt5dh\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051280 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f848653-564d-4a41-ab3b-3e509a42076e-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051289 4756 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051297 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051305 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051313 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d61f6dad-2b6f-475c-8516-127db0c852c2-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.051321 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f848653-564d-4a41-ab3b-3e509a42076e-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.052110 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-proxy-ca-bundles\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.053387 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-client-ca\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.054611 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-config\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.055536 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/637ecd17-93b8-41e7-83f4-210dc564be35-serving-cert\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.065907 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxcqc\" (UniqueName: \"kubernetes.io/projected/637ecd17-93b8-41e7-83f4-210dc564be35-kube-api-access-dxcqc\") pod \"controller-manager-68b4cbfd7f-m98m9\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.201057 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.238100 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-k6pzt"] Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.452191 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9"] Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.774096 4756 generic.go:334] "Generic (PLEG): container finished" podID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerID="57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e" exitCode=0 Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.774159 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw49c" event={"ID":"95a68714-9c47-4457-a37e-90a084d1a2b0","Type":"ContainerDied","Data":"57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.779599 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz2qm" event={"ID":"78ac0913-94cb-4f26-934d-bd005399d428","Type":"ContainerStarted","Data":"eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.781899 4756 generic.go:334] "Generic (PLEG): container finished" podID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerID="ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488" exitCode=0 Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.781952 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s64pq" event={"ID":"fc312654-25ec-4cf1-b35b-ea49aa157603","Type":"ContainerDied","Data":"ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.784338 4756 generic.go:334] "Generic (PLEG): container finished" podID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerID="37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32" exitCode=0 Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.784397 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frm8n" event={"ID":"cdb0e910-6e83-440e-bfb6-5e78e960c99d","Type":"ContainerDied","Data":"37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.788335 4756 generic.go:334] "Generic (PLEG): container finished" podID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerID="9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518" exitCode=0 Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.788379 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6bml" event={"ID":"775639e3-1ecc-4ccd-8207-fb6a557e49b3","Type":"ContainerDied","Data":"9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.790892 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" event={"ID":"717b7410-dd64-44cb-ba9b-3436d82ebb95","Type":"ContainerStarted","Data":"c3a8fdd75b81c189805ac6fd093ea62daf0c42c73bd4a630910d788354125d39"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.792168 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpmvh" event={"ID":"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37","Type":"ContainerStarted","Data":"8435a48cd0686e3b4948c6ac94eb3b44051c2437fe328c515f1f71aec627a68b"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.795923 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" event={"ID":"637ecd17-93b8-41e7-83f4-210dc564be35","Type":"ContainerStarted","Data":"3fd53fc2576962ab0c1d7b3fdd15fa7deb20f21403309edd1e15783b9fa0e181"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.801652 4756 generic.go:334] "Generic (PLEG): container finished" podID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerID="0a1d9e290ff51d498ca7c496c90f9dc03881786a14fadc97571aa3368ac6caa5" exitCode=0 Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.802025 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2lwz" event={"ID":"a22de99d-c3fa-444b-ae14-7e8de2145e01","Type":"ContainerDied","Data":"0a1d9e290ff51d498ca7c496c90f9dc03881786a14fadc97571aa3368ac6caa5"} Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.802133 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xn68m" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.802189 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q" Feb 03 09:13:07 crc kubenswrapper[4756]: E0203 09:13:07.807352 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zrjwc" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.893768 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xn68m"] Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.898156 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xn68m"] Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.943530 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q"] Feb 03 09:13:07 crc kubenswrapper[4756]: I0203 09:13:07.948855 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hzt8q"] Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.810272 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" event={"ID":"717b7410-dd64-44cb-ba9b-3436d82ebb95","Type":"ContainerStarted","Data":"04c88dc5352ebc5f1ecef5e5f6bdc68e78a9ee913e34611f830fc30327c516f1"} Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.812729 4756 generic.go:334] "Generic (PLEG): container finished" podID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerID="8435a48cd0686e3b4948c6ac94eb3b44051c2437fe328c515f1f71aec627a68b" exitCode=0 Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.812761 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpmvh" event={"ID":"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37","Type":"ContainerDied","Data":"8435a48cd0686e3b4948c6ac94eb3b44051c2437fe328c515f1f71aec627a68b"} Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.816192 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" event={"ID":"637ecd17-93b8-41e7-83f4-210dc564be35","Type":"ContainerStarted","Data":"d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4"} Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.816485 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.820401 4756 generic.go:334] "Generic (PLEG): container finished" podID="78ac0913-94cb-4f26-934d-bd005399d428" containerID="eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85" exitCode=0 Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.820437 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz2qm" event={"ID":"78ac0913-94cb-4f26-934d-bd005399d428","Type":"ContainerDied","Data":"eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85"} Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.826612 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:08 crc kubenswrapper[4756]: I0203 09:13:08.851730 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" podStartSLOduration=16.851710693 podStartE2EDuration="16.851710693s" podCreationTimestamp="2026-02-03 09:12:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:08.850230816 +0000 UTC m=+180.000698191" watchObservedRunningTime="2026-02-03 09:13:08.851710693 +0000 UTC m=+180.002178068" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.635156 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f848653-564d-4a41-ab3b-3e509a42076e" path="/var/lib/kubelet/pods/3f848653-564d-4a41-ab3b-3e509a42076e/volumes" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.636280 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d61f6dad-2b6f-475c-8516-127db0c852c2" path="/var/lib/kubelet/pods/d61f6dad-2b6f-475c-8516-127db0c852c2/volumes" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.733680 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8"] Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.734301 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.736059 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.737278 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.737473 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.737497 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.737573 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.737684 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.744963 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8"] Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.836195 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw49c" event={"ID":"95a68714-9c47-4457-a37e-90a084d1a2b0","Type":"ContainerStarted","Data":"7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109"} Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.838586 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k6pzt" event={"ID":"717b7410-dd64-44cb-ba9b-3436d82ebb95","Type":"ContainerStarted","Data":"baa4db543028d362cbc0f92237a0bd6debd94d91bc7787a88d9cf3b0eb23c494"} Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.852158 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xw49c" podStartSLOduration=2.8538389 podStartE2EDuration="33.852139128s" podCreationTimestamp="2026-02-03 09:12:36 +0000 UTC" firstStartedPulling="2026-02-03 09:12:38.314709984 +0000 UTC m=+149.465177359" lastFinishedPulling="2026-02-03 09:13:09.313010212 +0000 UTC m=+180.463477587" observedRunningTime="2026-02-03 09:13:09.852008763 +0000 UTC m=+181.002476138" watchObservedRunningTime="2026-02-03 09:13:09.852139128 +0000 UTC m=+181.002606503" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.868962 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-k6pzt" podStartSLOduration=161.868944228 podStartE2EDuration="2m41.868944228s" podCreationTimestamp="2026-02-03 09:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:09.867429431 +0000 UTC m=+181.017896806" watchObservedRunningTime="2026-02-03 09:13:09.868944228 +0000 UTC m=+181.019411603" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.888270 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-client-ca\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.888336 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5dhq\" (UniqueName: \"kubernetes.io/projected/cea9f381-e130-466c-93f7-3089b39d6e0f-kube-api-access-x5dhq\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.888357 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea9f381-e130-466c-93f7-3089b39d6e0f-serving-cert\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.888388 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-config\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.990232 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5dhq\" (UniqueName: \"kubernetes.io/projected/cea9f381-e130-466c-93f7-3089b39d6e0f-kube-api-access-x5dhq\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.990612 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea9f381-e130-466c-93f7-3089b39d6e0f-serving-cert\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.990666 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-config\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.990752 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-client-ca\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.991867 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-client-ca\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:09 crc kubenswrapper[4756]: I0203 09:13:09.992291 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-config\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.002617 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea9f381-e130-466c-93f7-3089b39d6e0f-serving-cert\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.009986 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5dhq\" (UniqueName: \"kubernetes.io/projected/cea9f381-e130-466c-93f7-3089b39d6e0f-kube-api-access-x5dhq\") pod \"route-controller-manager-7777b87969-dx2r8\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.055946 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.736302 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8"] Feb 03 09:13:10 crc kubenswrapper[4756]: W0203 09:13:10.746089 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcea9f381_e130_466c_93f7_3089b39d6e0f.slice/crio-82293fbf2c94904e47b9f705665598b6c877ec5b587b46aaa70452069e05b30a WatchSource:0}: Error finding container 82293fbf2c94904e47b9f705665598b6c877ec5b587b46aaa70452069e05b30a: Status 404 returned error can't find the container with id 82293fbf2c94904e47b9f705665598b6c877ec5b587b46aaa70452069e05b30a Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.849466 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s64pq" event={"ID":"fc312654-25ec-4cf1-b35b-ea49aa157603","Type":"ContainerStarted","Data":"9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239"} Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.853182 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frm8n" event={"ID":"cdb0e910-6e83-440e-bfb6-5e78e960c99d","Type":"ContainerStarted","Data":"885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907"} Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.859971 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpmvh" event={"ID":"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37","Type":"ContainerStarted","Data":"f1d033f80c3bc8ebade167bd6485df93ab1072ae42f3592ef7c6231cb95b6f5f"} Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.870408 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz2qm" event={"ID":"78ac0913-94cb-4f26-934d-bd005399d428","Type":"ContainerStarted","Data":"8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5"} Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.879359 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9sbkd" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.881687 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" event={"ID":"cea9f381-e130-466c-93f7-3089b39d6e0f","Type":"ContainerStarted","Data":"82293fbf2c94904e47b9f705665598b6c877ec5b587b46aaa70452069e05b30a"} Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.884066 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2lwz" event={"ID":"a22de99d-c3fa-444b-ae14-7e8de2145e01","Type":"ContainerStarted","Data":"fdf0815c1b935dca122eaa0702bdc3b4126539434c7cc4e6bb8a6e297947a608"} Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.900198 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qpmvh" podStartSLOduration=1.965597367 podStartE2EDuration="30.900176418s" podCreationTimestamp="2026-02-03 09:12:40 +0000 UTC" firstStartedPulling="2026-02-03 09:12:41.536009455 +0000 UTC m=+152.686476830" lastFinishedPulling="2026-02-03 09:13:10.470588506 +0000 UTC m=+181.621055881" observedRunningTime="2026-02-03 09:13:10.897133462 +0000 UTC m=+182.047600837" watchObservedRunningTime="2026-02-03 09:13:10.900176418 +0000 UTC m=+182.050643793" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.900382 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s64pq" podStartSLOduration=1.820145916 podStartE2EDuration="32.900378826s" podCreationTimestamp="2026-02-03 09:12:38 +0000 UTC" firstStartedPulling="2026-02-03 09:12:39.471780019 +0000 UTC m=+150.622247394" lastFinishedPulling="2026-02-03 09:13:10.552012929 +0000 UTC m=+181.702480304" observedRunningTime="2026-02-03 09:13:10.869408545 +0000 UTC m=+182.019875920" watchObservedRunningTime="2026-02-03 09:13:10.900378826 +0000 UTC m=+182.050846191" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.924541 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-frm8n" podStartSLOduration=3.082709 podStartE2EDuration="32.924523855s" podCreationTimestamp="2026-02-03 09:12:38 +0000 UTC" firstStartedPulling="2026-02-03 09:12:40.485525772 +0000 UTC m=+151.635993147" lastFinishedPulling="2026-02-03 09:13:10.327340627 +0000 UTC m=+181.477808002" observedRunningTime="2026-02-03 09:13:10.921734209 +0000 UTC m=+182.072201604" watchObservedRunningTime="2026-02-03 09:13:10.924523855 +0000 UTC m=+182.074991230" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.996287 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pz2qm" podStartSLOduration=2.9321117389999998 podStartE2EDuration="31.996253649s" podCreationTimestamp="2026-02-03 09:12:39 +0000 UTC" firstStartedPulling="2026-02-03 09:12:41.517550401 +0000 UTC m=+152.668017776" lastFinishedPulling="2026-02-03 09:13:10.581692311 +0000 UTC m=+181.732159686" observedRunningTime="2026-02-03 09:13:10.995851124 +0000 UTC m=+182.146318499" watchObservedRunningTime="2026-02-03 09:13:10.996253649 +0000 UTC m=+182.146721024" Feb 03 09:13:10 crc kubenswrapper[4756]: I0203 09:13:10.998243 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s2lwz" podStartSLOduration=4.032145875 podStartE2EDuration="34.998235085s" podCreationTimestamp="2026-02-03 09:12:36 +0000 UTC" firstStartedPulling="2026-02-03 09:12:39.470032003 +0000 UTC m=+150.620499378" lastFinishedPulling="2026-02-03 09:13:10.436121213 +0000 UTC m=+181.586588588" observedRunningTime="2026-02-03 09:13:10.973855665 +0000 UTC m=+182.124323040" watchObservedRunningTime="2026-02-03 09:13:10.998235085 +0000 UTC m=+182.148702460" Feb 03 09:13:11 crc kubenswrapper[4756]: I0203 09:13:11.891159 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" event={"ID":"cea9f381-e130-466c-93f7-3089b39d6e0f","Type":"ContainerStarted","Data":"2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a"} Feb 03 09:13:11 crc kubenswrapper[4756]: I0203 09:13:11.894969 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:11 crc kubenswrapper[4756]: I0203 09:13:11.898611 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6bml" event={"ID":"775639e3-1ecc-4ccd-8207-fb6a557e49b3","Type":"ContainerStarted","Data":"d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55"} Feb 03 09:13:11 crc kubenswrapper[4756]: I0203 09:13:11.902043 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:11 crc kubenswrapper[4756]: I0203 09:13:11.911528 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" podStartSLOduration=19.911483148 podStartE2EDuration="19.911483148s" podCreationTimestamp="2026-02-03 09:12:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:11.908883719 +0000 UTC m=+183.059351094" watchObservedRunningTime="2026-02-03 09:13:11.911483148 +0000 UTC m=+183.061950543" Feb 03 09:13:11 crc kubenswrapper[4756]: I0203 09:13:11.951643 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v6bml" podStartSLOduration=4.459359434 podStartE2EDuration="35.951619317s" podCreationTimestamp="2026-02-03 09:12:36 +0000 UTC" firstStartedPulling="2026-02-03 09:12:39.44821123 +0000 UTC m=+150.598678605" lastFinishedPulling="2026-02-03 09:13:10.940471113 +0000 UTC m=+182.090938488" observedRunningTime="2026-02-03 09:13:11.949089241 +0000 UTC m=+183.099556626" watchObservedRunningTime="2026-02-03 09:13:11.951619317 +0000 UTC m=+183.102086702" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.049597 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9"] Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.050083 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" podUID="637ecd17-93b8-41e7-83f4-210dc564be35" containerName="controller-manager" containerID="cri-o://d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4" gracePeriod=30 Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.105460 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8"] Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.636644 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.723364 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/637ecd17-93b8-41e7-83f4-210dc564be35-serving-cert\") pod \"637ecd17-93b8-41e7-83f4-210dc564be35\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.723456 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxcqc\" (UniqueName: \"kubernetes.io/projected/637ecd17-93b8-41e7-83f4-210dc564be35-kube-api-access-dxcqc\") pod \"637ecd17-93b8-41e7-83f4-210dc564be35\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.723493 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-client-ca\") pod \"637ecd17-93b8-41e7-83f4-210dc564be35\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.723543 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-config\") pod \"637ecd17-93b8-41e7-83f4-210dc564be35\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.723597 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-proxy-ca-bundles\") pod \"637ecd17-93b8-41e7-83f4-210dc564be35\" (UID: \"637ecd17-93b8-41e7-83f4-210dc564be35\") " Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.724152 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-client-ca" (OuterVolumeSpecName: "client-ca") pod "637ecd17-93b8-41e7-83f4-210dc564be35" (UID: "637ecd17-93b8-41e7-83f4-210dc564be35"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.724299 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "637ecd17-93b8-41e7-83f4-210dc564be35" (UID: "637ecd17-93b8-41e7-83f4-210dc564be35"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.724411 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-config" (OuterVolumeSpecName: "config") pod "637ecd17-93b8-41e7-83f4-210dc564be35" (UID: "637ecd17-93b8-41e7-83f4-210dc564be35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.729428 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/637ecd17-93b8-41e7-83f4-210dc564be35-kube-api-access-dxcqc" (OuterVolumeSpecName: "kube-api-access-dxcqc") pod "637ecd17-93b8-41e7-83f4-210dc564be35" (UID: "637ecd17-93b8-41e7-83f4-210dc564be35"). InnerVolumeSpecName "kube-api-access-dxcqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.737498 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/637ecd17-93b8-41e7-83f4-210dc564be35-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "637ecd17-93b8-41e7-83f4-210dc564be35" (UID: "637ecd17-93b8-41e7-83f4-210dc564be35"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.824969 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/637ecd17-93b8-41e7-83f4-210dc564be35-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.825033 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxcqc\" (UniqueName: \"kubernetes.io/projected/637ecd17-93b8-41e7-83f4-210dc564be35-kube-api-access-dxcqc\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.829509 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.830403 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.831658 4756 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/637ecd17-93b8-41e7-83f4-210dc564be35-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.911023 4756 generic.go:334] "Generic (PLEG): container finished" podID="637ecd17-93b8-41e7-83f4-210dc564be35" containerID="d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4" exitCode=0 Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.911129 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" event={"ID":"637ecd17-93b8-41e7-83f4-210dc564be35","Type":"ContainerDied","Data":"d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4"} Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.911176 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" event={"ID":"637ecd17-93b8-41e7-83f4-210dc564be35","Type":"ContainerDied","Data":"3fd53fc2576962ab0c1d7b3fdd15fa7deb20f21403309edd1e15783b9fa0e181"} Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.911194 4756 scope.go:117] "RemoveContainer" containerID="d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.911198 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.926840 4756 scope.go:117] "RemoveContainer" containerID="d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4" Feb 03 09:13:12 crc kubenswrapper[4756]: E0203 09:13:12.928865 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4\": container with ID starting with d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4 not found: ID does not exist" containerID="d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.928907 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4"} err="failed to get container status \"d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4\": rpc error: code = NotFound desc = could not find container \"d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4\": container with ID starting with d182c95eb259e8dfb790a9c60220426dd9f9fa4892cfe15a6472cb92bf6b46f4 not found: ID does not exist" Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.949633 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9"] Feb 03 09:13:12 crc kubenswrapper[4756]: I0203 09:13:12.953167 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-68b4cbfd7f-m98m9"] Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.566712 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.566810 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.620775 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="637ecd17-93b8-41e7-83f4-210dc564be35" path="/var/lib/kubelet/pods/637ecd17-93b8-41e7-83f4-210dc564be35/volumes" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.739545 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7784f78584-xhkgl"] Feb 03 09:13:13 crc kubenswrapper[4756]: E0203 09:13:13.739776 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637ecd17-93b8-41e7-83f4-210dc564be35" containerName="controller-manager" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.739794 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="637ecd17-93b8-41e7-83f4-210dc564be35" containerName="controller-manager" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.739940 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="637ecd17-93b8-41e7-83f4-210dc564be35" containerName="controller-manager" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.740405 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.743672 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.744284 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.744597 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.744714 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.745322 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.746940 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.751309 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.756645 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7784f78584-xhkgl"] Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.843294 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tdbp\" (UniqueName: \"kubernetes.io/projected/4c0f8247-0f03-42e2-b123-707431218ab6-kube-api-access-6tdbp\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.843360 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c0f8247-0f03-42e2-b123-707431218ab6-serving-cert\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.843386 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-proxy-ca-bundles\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.843408 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-config\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.843453 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-client-ca\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.928739 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" podUID="cea9f381-e130-466c-93f7-3089b39d6e0f" containerName="route-controller-manager" containerID="cri-o://2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a" gracePeriod=30 Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.944568 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tdbp\" (UniqueName: \"kubernetes.io/projected/4c0f8247-0f03-42e2-b123-707431218ab6-kube-api-access-6tdbp\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.944623 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c0f8247-0f03-42e2-b123-707431218ab6-serving-cert\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.944642 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-proxy-ca-bundles\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.944668 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-config\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.944697 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-client-ca\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.945531 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-client-ca\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.947083 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-proxy-ca-bundles\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.948031 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-config\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.950719 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c0f8247-0f03-42e2-b123-707431218ab6-serving-cert\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:13 crc kubenswrapper[4756]: I0203 09:13:13.964109 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tdbp\" (UniqueName: \"kubernetes.io/projected/4c0f8247-0f03-42e2-b123-707431218ab6-kube-api-access-6tdbp\") pod \"controller-manager-7784f78584-xhkgl\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.103987 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.282028 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.349476 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-config\") pod \"cea9f381-e130-466c-93f7-3089b39d6e0f\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.349719 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5dhq\" (UniqueName: \"kubernetes.io/projected/cea9f381-e130-466c-93f7-3089b39d6e0f-kube-api-access-x5dhq\") pod \"cea9f381-e130-466c-93f7-3089b39d6e0f\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.349794 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea9f381-e130-466c-93f7-3089b39d6e0f-serving-cert\") pod \"cea9f381-e130-466c-93f7-3089b39d6e0f\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.349848 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-client-ca\") pod \"cea9f381-e130-466c-93f7-3089b39d6e0f\" (UID: \"cea9f381-e130-466c-93f7-3089b39d6e0f\") " Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.350138 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-config" (OuterVolumeSpecName: "config") pod "cea9f381-e130-466c-93f7-3089b39d6e0f" (UID: "cea9f381-e130-466c-93f7-3089b39d6e0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.350272 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-client-ca" (OuterVolumeSpecName: "client-ca") pod "cea9f381-e130-466c-93f7-3089b39d6e0f" (UID: "cea9f381-e130-466c-93f7-3089b39d6e0f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.350481 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.350497 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cea9f381-e130-466c-93f7-3089b39d6e0f-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.353241 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cea9f381-e130-466c-93f7-3089b39d6e0f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cea9f381-e130-466c-93f7-3089b39d6e0f" (UID: "cea9f381-e130-466c-93f7-3089b39d6e0f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.353579 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cea9f381-e130-466c-93f7-3089b39d6e0f-kube-api-access-x5dhq" (OuterVolumeSpecName: "kube-api-access-x5dhq") pod "cea9f381-e130-466c-93f7-3089b39d6e0f" (UID: "cea9f381-e130-466c-93f7-3089b39d6e0f"). InnerVolumeSpecName "kube-api-access-x5dhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.451373 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5dhq\" (UniqueName: \"kubernetes.io/projected/cea9f381-e130-466c-93f7-3089b39d6e0f-kube-api-access-x5dhq\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.451416 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cea9f381-e130-466c-93f7-3089b39d6e0f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.564626 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7784f78584-xhkgl"] Feb 03 09:13:14 crc kubenswrapper[4756]: W0203 09:13:14.568569 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c0f8247_0f03_42e2_b123_707431218ab6.slice/crio-18da30e507446c4fa66f2c775415214ee2ec23061dc8c57d1ecdaa69c22e50c2 WatchSource:0}: Error finding container 18da30e507446c4fa66f2c775415214ee2ec23061dc8c57d1ecdaa69c22e50c2: Status 404 returned error can't find the container with id 18da30e507446c4fa66f2c775415214ee2ec23061dc8c57d1ecdaa69c22e50c2 Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.934731 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" event={"ID":"4c0f8247-0f03-42e2-b123-707431218ab6","Type":"ContainerStarted","Data":"0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00"} Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.935437 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.935513 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" event={"ID":"4c0f8247-0f03-42e2-b123-707431218ab6","Type":"ContainerStarted","Data":"18da30e507446c4fa66f2c775415214ee2ec23061dc8c57d1ecdaa69c22e50c2"} Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.937198 4756 generic.go:334] "Generic (PLEG): container finished" podID="cea9f381-e130-466c-93f7-3089b39d6e0f" containerID="2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a" exitCode=0 Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.937261 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" event={"ID":"cea9f381-e130-466c-93f7-3089b39d6e0f","Type":"ContainerDied","Data":"2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a"} Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.937284 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" event={"ID":"cea9f381-e130-466c-93f7-3089b39d6e0f","Type":"ContainerDied","Data":"82293fbf2c94904e47b9f705665598b6c877ec5b587b46aaa70452069e05b30a"} Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.937328 4756 scope.go:117] "RemoveContainer" containerID="2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.937493 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.940144 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.950357 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" podStartSLOduration=2.950338356 podStartE2EDuration="2.950338356s" podCreationTimestamp="2026-02-03 09:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:14.94886241 +0000 UTC m=+186.099329805" watchObservedRunningTime="2026-02-03 09:13:14.950338356 +0000 UTC m=+186.100805731" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.952235 4756 scope.go:117] "RemoveContainer" containerID="2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a" Feb 03 09:13:14 crc kubenswrapper[4756]: E0203 09:13:14.952893 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a\": container with ID starting with 2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a not found: ID does not exist" containerID="2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a" Feb 03 09:13:14 crc kubenswrapper[4756]: I0203 09:13:14.952933 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a"} err="failed to get container status \"2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a\": rpc error: code = NotFound desc = could not find container \"2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a\": container with ID starting with 2943de501f68fcef0a3b15753b5fa874488d26ecf62f757c46d5ce87dab60a0a not found: ID does not exist" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.018110 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8"] Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.048599 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7777b87969-dx2r8"] Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.620561 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cea9f381-e130-466c-93f7-3089b39d6e0f" path="/var/lib/kubelet/pods/cea9f381-e130-466c-93f7-3089b39d6e0f/volumes" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.737605 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv"] Feb 03 09:13:15 crc kubenswrapper[4756]: E0203 09:13:15.737853 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea9f381-e130-466c-93f7-3089b39d6e0f" containerName="route-controller-manager" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.737867 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea9f381-e130-466c-93f7-3089b39d6e0f" containerName="route-controller-manager" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.739527 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="cea9f381-e130-466c-93f7-3089b39d6e0f" containerName="route-controller-manager" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.740124 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.742053 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.742672 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.742695 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.742914 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.742953 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.743079 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.754202 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv"] Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.868969 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-config\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.869105 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-client-ca\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.869150 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5cf300-5392-441c-9cb2-9a7b0268ff1d-serving-cert\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.869217 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjxrs\" (UniqueName: \"kubernetes.io/projected/da5cf300-5392-441c-9cb2-9a7b0268ff1d-kube-api-access-kjxrs\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.969920 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjxrs\" (UniqueName: \"kubernetes.io/projected/da5cf300-5392-441c-9cb2-9a7b0268ff1d-kube-api-access-kjxrs\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.969981 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-config\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.970024 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-client-ca\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.970045 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5cf300-5392-441c-9cb2-9a7b0268ff1d-serving-cert\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.970986 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-client-ca\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.971807 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-config\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.978322 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5cf300-5392-441c-9cb2-9a7b0268ff1d-serving-cert\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:15 crc kubenswrapper[4756]: I0203 09:13:15.989470 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjxrs\" (UniqueName: \"kubernetes.io/projected/da5cf300-5392-441c-9cb2-9a7b0268ff1d-kube-api-access-kjxrs\") pod \"route-controller-manager-69c8657f57-q6fpv\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:16 crc kubenswrapper[4756]: I0203 09:13:16.054494 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:16 crc kubenswrapper[4756]: I0203 09:13:16.451740 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv"] Feb 03 09:13:16 crc kubenswrapper[4756]: I0203 09:13:16.547356 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 03 09:13:16 crc kubenswrapper[4756]: I0203 09:13:16.959518 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" event={"ID":"da5cf300-5392-441c-9cb2-9a7b0268ff1d","Type":"ContainerStarted","Data":"e17783777ff6d66013c937662becb067dde567999d3d7bb403273ecfcc374bf7"} Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.030093 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.030153 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.151178 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.151315 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.346661 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.346723 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.485861 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.489085 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.489605 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.930736 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8jhbb"] Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.965530 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" event={"ID":"da5cf300-5392-441c-9cb2-9a7b0268ff1d","Type":"ContainerStarted","Data":"6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394"} Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.966536 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:17 crc kubenswrapper[4756]: I0203 09:13:17.972572 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.016771 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" podStartSLOduration=6.016749094 podStartE2EDuration="6.016749094s" podCreationTimestamp="2026-02-03 09:13:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:18.012933518 +0000 UTC m=+189.163400903" watchObservedRunningTime="2026-02-03 09:13:18.016749094 +0000 UTC m=+189.167216469" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.102641 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.111793 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.118068 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.626311 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.626946 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.629092 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.633231 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.641890 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.704176 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.704258 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.805504 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.805591 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.805686 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.826948 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.911827 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.912096 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.955116 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:13:18 crc kubenswrapper[4756]: I0203 09:13:18.976045 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:19 crc kubenswrapper[4756]: I0203 09:13:19.019755 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:13:19 crc kubenswrapper[4756]: I0203 09:13:19.114156 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s2lwz"] Feb 03 09:13:19 crc kubenswrapper[4756]: I0203 09:13:19.268211 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:13:19 crc kubenswrapper[4756]: I0203 09:13:19.268269 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:13:19 crc kubenswrapper[4756]: I0203 09:13:19.316164 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:13:19 crc kubenswrapper[4756]: I0203 09:13:19.400781 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 03 09:13:19 crc kubenswrapper[4756]: W0203 09:13:19.407872 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod316819ba_fb6c_4e74_ae0e_7f08aa5a2a5c.slice/crio-99a5668fdf6b0e40f38b206b0f99c74d09fb8969df7c7980595e9bf3bfe2b341 WatchSource:0}: Error finding container 99a5668fdf6b0e40f38b206b0f99c74d09fb8969df7c7980595e9bf3bfe2b341: Status 404 returned error can't find the container with id 99a5668fdf6b0e40f38b206b0f99c74d09fb8969df7c7980595e9bf3bfe2b341 Feb 03 09:13:19 crc kubenswrapper[4756]: I0203 09:13:19.975459 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c","Type":"ContainerStarted","Data":"99a5668fdf6b0e40f38b206b0f99c74d09fb8969df7c7980595e9bf3bfe2b341"} Feb 03 09:13:19 crc kubenswrapper[4756]: I0203 09:13:19.975918 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s2lwz" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerName="registry-server" containerID="cri-o://fdf0815c1b935dca122eaa0702bdc3b4126539434c7cc4e6bb8a6e297947a608" gracePeriod=2 Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.063964 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.218677 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.219031 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.269434 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.643264 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.643310 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.685926 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.996137 4756 generic.go:334] "Generic (PLEG): container finished" podID="316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c" containerID="66c2eb74da610afe70f3aba522123236b88d42f0cb9ac1d976e7a62772f28a8f" exitCode=0 Feb 03 09:13:20 crc kubenswrapper[4756]: I0203 09:13:20.996248 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c","Type":"ContainerDied","Data":"66c2eb74da610afe70f3aba522123236b88d42f0cb9ac1d976e7a62772f28a8f"} Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.003569 4756 generic.go:334] "Generic (PLEG): container finished" podID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerID="fdf0815c1b935dca122eaa0702bdc3b4126539434c7cc4e6bb8a6e297947a608" exitCode=0 Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.003621 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2lwz" event={"ID":"a22de99d-c3fa-444b-ae14-7e8de2145e01","Type":"ContainerDied","Data":"fdf0815c1b935dca122eaa0702bdc3b4126539434c7cc4e6bb8a6e297947a608"} Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.060382 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.064198 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.161821 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.251608 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-catalog-content\") pod \"a22de99d-c3fa-444b-ae14-7e8de2145e01\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.251713 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm8kw\" (UniqueName: \"kubernetes.io/projected/a22de99d-c3fa-444b-ae14-7e8de2145e01-kube-api-access-wm8kw\") pod \"a22de99d-c3fa-444b-ae14-7e8de2145e01\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.251789 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-utilities\") pod \"a22de99d-c3fa-444b-ae14-7e8de2145e01\" (UID: \"a22de99d-c3fa-444b-ae14-7e8de2145e01\") " Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.252804 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-utilities" (OuterVolumeSpecName: "utilities") pod "a22de99d-c3fa-444b-ae14-7e8de2145e01" (UID: "a22de99d-c3fa-444b-ae14-7e8de2145e01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.262643 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a22de99d-c3fa-444b-ae14-7e8de2145e01-kube-api-access-wm8kw" (OuterVolumeSpecName: "kube-api-access-wm8kw") pod "a22de99d-c3fa-444b-ae14-7e8de2145e01" (UID: "a22de99d-c3fa-444b-ae14-7e8de2145e01"). InnerVolumeSpecName "kube-api-access-wm8kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.306166 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a22de99d-c3fa-444b-ae14-7e8de2145e01" (UID: "a22de99d-c3fa-444b-ae14-7e8de2145e01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.360050 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.360090 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm8kw\" (UniqueName: \"kubernetes.io/projected/a22de99d-c3fa-444b-ae14-7e8de2145e01-kube-api-access-wm8kw\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.360100 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a22de99d-c3fa-444b-ae14-7e8de2145e01-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:21 crc kubenswrapper[4756]: I0203 09:13:21.514118 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-frm8n"] Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.011861 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrjwc" event={"ID":"a933c882-e11d-46a8-8f67-f77ebc97367c","Type":"ContainerStarted","Data":"86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab"} Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.017422 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2lwz" event={"ID":"a22de99d-c3fa-444b-ae14-7e8de2145e01","Type":"ContainerDied","Data":"877fb99d49a261ca592ecd832701086142134dc2eb632deb97aae3396c81a42f"} Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.017502 4756 scope.go:117] "RemoveContainer" containerID="fdf0815c1b935dca122eaa0702bdc3b4126539434c7cc4e6bb8a6e297947a608" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.017990 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-frm8n" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerName="registry-server" containerID="cri-o://885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907" gracePeriod=2 Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.018138 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2lwz" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.032557 4756 scope.go:117] "RemoveContainer" containerID="0a1d9e290ff51d498ca7c496c90f9dc03881786a14fadc97571aa3368ac6caa5" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.070181 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s2lwz"] Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.070355 4756 scope.go:117] "RemoveContainer" containerID="ec1737f649380811b51c94caa9ea9e6b1c07426d6c0e12da12284e8faa6443e8" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.076409 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s2lwz"] Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.389101 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.477266 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kube-api-access\") pod \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\" (UID: \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\") " Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.477865 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kubelet-dir\") pod \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\" (UID: \"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c\") " Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.478155 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c" (UID: "316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.483714 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c" (UID: "316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.579085 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.579131 4756 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.630350 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.781797 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndnt8\" (UniqueName: \"kubernetes.io/projected/cdb0e910-6e83-440e-bfb6-5e78e960c99d-kube-api-access-ndnt8\") pod \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.781893 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-catalog-content\") pod \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.782027 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-utilities\") pod \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\" (UID: \"cdb0e910-6e83-440e-bfb6-5e78e960c99d\") " Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.782780 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-utilities" (OuterVolumeSpecName: "utilities") pod "cdb0e910-6e83-440e-bfb6-5e78e960c99d" (UID: "cdb0e910-6e83-440e-bfb6-5e78e960c99d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.782984 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.785380 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdb0e910-6e83-440e-bfb6-5e78e960c99d-kube-api-access-ndnt8" (OuterVolumeSpecName: "kube-api-access-ndnt8") pod "cdb0e910-6e83-440e-bfb6-5e78e960c99d" (UID: "cdb0e910-6e83-440e-bfb6-5e78e960c99d"). InnerVolumeSpecName "kube-api-access-ndnt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.810076 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cdb0e910-6e83-440e-bfb6-5e78e960c99d" (UID: "cdb0e910-6e83-440e-bfb6-5e78e960c99d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.884541 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndnt8\" (UniqueName: \"kubernetes.io/projected/cdb0e910-6e83-440e-bfb6-5e78e960c99d-kube-api-access-ndnt8\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:22 crc kubenswrapper[4756]: I0203 09:13:22.884577 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdb0e910-6e83-440e-bfb6-5e78e960c99d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.024647 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.024639 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c","Type":"ContainerDied","Data":"99a5668fdf6b0e40f38b206b0f99c74d09fb8969df7c7980595e9bf3bfe2b341"} Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.024768 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99a5668fdf6b0e40f38b206b0f99c74d09fb8969df7c7980595e9bf3bfe2b341" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.027478 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrjwc" event={"ID":"a933c882-e11d-46a8-8f67-f77ebc97367c","Type":"ContainerDied","Data":"86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab"} Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.027435 4756 generic.go:334] "Generic (PLEG): container finished" podID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerID="86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab" exitCode=0 Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.033155 4756 generic.go:334] "Generic (PLEG): container finished" podID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerID="885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907" exitCode=0 Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.033385 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frm8n" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.033187 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frm8n" event={"ID":"cdb0e910-6e83-440e-bfb6-5e78e960c99d","Type":"ContainerDied","Data":"885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907"} Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.042618 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frm8n" event={"ID":"cdb0e910-6e83-440e-bfb6-5e78e960c99d","Type":"ContainerDied","Data":"2aabdeccb55adc8a6cf051e250d633cfd6de3c59b15e36a0f7be641358b0e478"} Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.042664 4756 scope.go:117] "RemoveContainer" containerID="885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.071960 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-frm8n"] Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.074671 4756 scope.go:117] "RemoveContainer" containerID="37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.075100 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-frm8n"] Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.088599 4756 scope.go:117] "RemoveContainer" containerID="93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.103089 4756 scope.go:117] "RemoveContainer" containerID="885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907" Feb 03 09:13:23 crc kubenswrapper[4756]: E0203 09:13:23.105311 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907\": container with ID starting with 885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907 not found: ID does not exist" containerID="885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.105352 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907"} err="failed to get container status \"885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907\": rpc error: code = NotFound desc = could not find container \"885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907\": container with ID starting with 885369b1967ecf11950ff04488bf60f2908eb1c9bdd4b56b8f3419f5f206e907 not found: ID does not exist" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.105384 4756 scope.go:117] "RemoveContainer" containerID="37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32" Feb 03 09:13:23 crc kubenswrapper[4756]: E0203 09:13:23.105829 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32\": container with ID starting with 37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32 not found: ID does not exist" containerID="37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.105853 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32"} err="failed to get container status \"37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32\": rpc error: code = NotFound desc = could not find container \"37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32\": container with ID starting with 37e9351cb238d4254df0274d332106be981ec577edbf3225efbfe1e18bbe1c32 not found: ID does not exist" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.105868 4756 scope.go:117] "RemoveContainer" containerID="93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd" Feb 03 09:13:23 crc kubenswrapper[4756]: E0203 09:13:23.106174 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd\": container with ID starting with 93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd not found: ID does not exist" containerID="93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.106198 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd"} err="failed to get container status \"93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd\": rpc error: code = NotFound desc = could not find container \"93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd\": container with ID starting with 93ebe7c8d002bdb768c56035f574ac3d51e6a84c840712bf66735c7b7d01e3bd not found: ID does not exist" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.620908 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" path="/var/lib/kubelet/pods/a22de99d-c3fa-444b-ae14-7e8de2145e01/volumes" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.621808 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" path="/var/lib/kubelet/pods/cdb0e910-6e83-440e-bfb6-5e78e960c99d/volumes" Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.913092 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qpmvh"] Feb 03 09:13:23 crc kubenswrapper[4756]: I0203 09:13:23.913317 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qpmvh" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerName="registry-server" containerID="cri-o://f1d033f80c3bc8ebade167bd6485df93ab1072ae42f3592ef7c6231cb95b6f5f" gracePeriod=2 Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.045837 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrjwc" event={"ID":"a933c882-e11d-46a8-8f67-f77ebc97367c","Type":"ContainerStarted","Data":"fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd"} Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.048892 4756 generic.go:334] "Generic (PLEG): container finished" podID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerID="f1d033f80c3bc8ebade167bd6485df93ab1072ae42f3592ef7c6231cb95b6f5f" exitCode=0 Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.048938 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpmvh" event={"ID":"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37","Type":"ContainerDied","Data":"f1d033f80c3bc8ebade167bd6485df93ab1072ae42f3592ef7c6231cb95b6f5f"} Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.066203 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zrjwc" podStartSLOduration=3.005782169 podStartE2EDuration="47.066181223s" podCreationTimestamp="2026-02-03 09:12:37 +0000 UTC" firstStartedPulling="2026-02-03 09:12:39.448588695 +0000 UTC m=+150.599056070" lastFinishedPulling="2026-02-03 09:13:23.508987749 +0000 UTC m=+194.659455124" observedRunningTime="2026-02-03 09:13:24.064918935 +0000 UTC m=+195.215386320" watchObservedRunningTime="2026-02-03 09:13:24.066181223 +0000 UTC m=+195.216648598" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.417438 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.544609 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-catalog-content\") pod \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.545170 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trd5n\" (UniqueName: \"kubernetes.io/projected/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-kube-api-access-trd5n\") pod \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.545207 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-utilities\") pod \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\" (UID: \"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37\") " Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.546077 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-utilities" (OuterVolumeSpecName: "utilities") pod "a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" (UID: "a6d20a8f-7b6e-4308-b951-9b29cd8c1b37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.552678 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-kube-api-access-trd5n" (OuterVolumeSpecName: "kube-api-access-trd5n") pod "a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" (UID: "a6d20a8f-7b6e-4308-b951-9b29cd8c1b37"). InnerVolumeSpecName "kube-api-access-trd5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.626832 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627076 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerName="extract-utilities" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627093 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerName="extract-utilities" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627107 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerName="extract-utilities" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627115 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerName="extract-utilities" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627124 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerName="extract-content" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627133 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerName="extract-content" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627140 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerName="extract-content" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627148 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerName="extract-content" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627160 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerName="extract-utilities" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627168 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerName="extract-utilities" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627181 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627188 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627203 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627210 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627220 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627228 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627240 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c" containerName="pruner" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627247 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c" containerName="pruner" Feb 03 09:13:24 crc kubenswrapper[4756]: E0203 09:13:24.627255 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerName="extract-content" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627262 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerName="extract-content" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627395 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22de99d-c3fa-444b-ae14-7e8de2145e01" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627408 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdb0e910-6e83-440e-bfb6-5e78e960c99d" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627420 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="316819ba-fb6c-4e74-ae0e-7f08aa5a2a5c" containerName="pruner" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627434 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" containerName="registry-server" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.627941 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.630880 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.630958 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.640271 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.647104 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trd5n\" (UniqueName: \"kubernetes.io/projected/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-kube-api-access-trd5n\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.648097 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.697693 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" (UID: "a6d20a8f-7b6e-4308-b951-9b29cd8c1b37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.748897 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-var-lock\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.748947 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6bdec88-abf9-41e5-832e-a3c52065622f-kube-api-access\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.749065 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.749218 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.850768 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6bdec88-abf9-41e5-832e-a3c52065622f-kube-api-access\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.850876 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.850944 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-var-lock\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.851031 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.851054 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-var-lock\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.868370 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6bdec88-abf9-41e5-832e-a3c52065622f-kube-api-access\") pod \"installer-9-crc\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:24 crc kubenswrapper[4756]: I0203 09:13:24.946984 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.057748 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpmvh" event={"ID":"a6d20a8f-7b6e-4308-b951-9b29cd8c1b37","Type":"ContainerDied","Data":"a5dcd811d66ce87825a7413c5e9f343716a305c64854bc27023683cb1670793e"} Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.057825 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpmvh" Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.057866 4756 scope.go:117] "RemoveContainer" containerID="f1d033f80c3bc8ebade167bd6485df93ab1072ae42f3592ef7c6231cb95b6f5f" Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.076806 4756 scope.go:117] "RemoveContainer" containerID="8435a48cd0686e3b4948c6ac94eb3b44051c2437fe328c515f1f71aec627a68b" Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.088305 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qpmvh"] Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.092153 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qpmvh"] Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.107480 4756 scope.go:117] "RemoveContainer" containerID="6a92eb4bbe0bd609976e45303572866ca0751213d2da93649e90354eb7e3291e" Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.362454 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 03 09:13:25 crc kubenswrapper[4756]: W0203 09:13:25.365703 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda6bdec88_abf9_41e5_832e_a3c52065622f.slice/crio-f3b191b0c3242af4350843c36921a19a15814a65da51221ad16cf275fbd08f54 WatchSource:0}: Error finding container f3b191b0c3242af4350843c36921a19a15814a65da51221ad16cf275fbd08f54: Status 404 returned error can't find the container with id f3b191b0c3242af4350843c36921a19a15814a65da51221ad16cf275fbd08f54 Feb 03 09:13:25 crc kubenswrapper[4756]: I0203 09:13:25.622470 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6d20a8f-7b6e-4308-b951-9b29cd8c1b37" path="/var/lib/kubelet/pods/a6d20a8f-7b6e-4308-b951-9b29cd8c1b37/volumes" Feb 03 09:13:26 crc kubenswrapper[4756]: I0203 09:13:26.066622 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a6bdec88-abf9-41e5-832e-a3c52065622f","Type":"ContainerStarted","Data":"2ed3bf2bf9f21ab9da027c864b0b2a206898a3da9631f46aff677cce9097317e"} Feb 03 09:13:26 crc kubenswrapper[4756]: I0203 09:13:26.066667 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a6bdec88-abf9-41e5-832e-a3c52065622f","Type":"ContainerStarted","Data":"f3b191b0c3242af4350843c36921a19a15814a65da51221ad16cf275fbd08f54"} Feb 03 09:13:26 crc kubenswrapper[4756]: I0203 09:13:26.088312 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.088291408 podStartE2EDuration="2.088291408s" podCreationTimestamp="2026-02-03 09:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:26.08058231 +0000 UTC m=+197.231049695" watchObservedRunningTime="2026-02-03 09:13:26.088291408 +0000 UTC m=+197.238758783" Feb 03 09:13:27 crc kubenswrapper[4756]: I0203 09:13:27.507259 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:13:27 crc kubenswrapper[4756]: I0203 09:13:27.507665 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:13:27 crc kubenswrapper[4756]: I0203 09:13:27.551024 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:13:28 crc kubenswrapper[4756]: I0203 09:13:28.121438 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.001657 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7784f78584-xhkgl"] Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.002257 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" podUID="4c0f8247-0f03-42e2-b123-707431218ab6" containerName="controller-manager" containerID="cri-o://0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00" gracePeriod=30 Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.018072 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv"] Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.018299 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" podUID="da5cf300-5392-441c-9cb2-9a7b0268ff1d" containerName="route-controller-manager" containerID="cri-o://6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394" gracePeriod=30 Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.349484 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zrjwc"] Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.350023 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zrjwc" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerName="registry-server" containerID="cri-o://fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd" gracePeriod=2 Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.671645 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.691928 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754584 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-config\") pod \"4c0f8247-0f03-42e2-b123-707431218ab6\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754625 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c0f8247-0f03-42e2-b123-707431218ab6-serving-cert\") pod \"4c0f8247-0f03-42e2-b123-707431218ab6\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754650 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-proxy-ca-bundles\") pod \"4c0f8247-0f03-42e2-b123-707431218ab6\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754667 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjxrs\" (UniqueName: \"kubernetes.io/projected/da5cf300-5392-441c-9cb2-9a7b0268ff1d-kube-api-access-kjxrs\") pod \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754697 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-client-ca\") pod \"4c0f8247-0f03-42e2-b123-707431218ab6\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754723 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5cf300-5392-441c-9cb2-9a7b0268ff1d-serving-cert\") pod \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754743 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-client-ca\") pod \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754784 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-config\") pod \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\" (UID: \"da5cf300-5392-441c-9cb2-9a7b0268ff1d\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.754841 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tdbp\" (UniqueName: \"kubernetes.io/projected/4c0f8247-0f03-42e2-b123-707431218ab6-kube-api-access-6tdbp\") pod \"4c0f8247-0f03-42e2-b123-707431218ab6\" (UID: \"4c0f8247-0f03-42e2-b123-707431218ab6\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.758550 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-client-ca" (OuterVolumeSpecName: "client-ca") pod "da5cf300-5392-441c-9cb2-9a7b0268ff1d" (UID: "da5cf300-5392-441c-9cb2-9a7b0268ff1d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.758955 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-client-ca" (OuterVolumeSpecName: "client-ca") pod "4c0f8247-0f03-42e2-b123-707431218ab6" (UID: "4c0f8247-0f03-42e2-b123-707431218ab6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.759047 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-config" (OuterVolumeSpecName: "config") pod "da5cf300-5392-441c-9cb2-9a7b0268ff1d" (UID: "da5cf300-5392-441c-9cb2-9a7b0268ff1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.759481 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-config" (OuterVolumeSpecName: "config") pod "4c0f8247-0f03-42e2-b123-707431218ab6" (UID: "4c0f8247-0f03-42e2-b123-707431218ab6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.761664 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da5cf300-5392-441c-9cb2-9a7b0268ff1d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "da5cf300-5392-441c-9cb2-9a7b0268ff1d" (UID: "da5cf300-5392-441c-9cb2-9a7b0268ff1d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.761703 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c0f8247-0f03-42e2-b123-707431218ab6-kube-api-access-6tdbp" (OuterVolumeSpecName: "kube-api-access-6tdbp") pod "4c0f8247-0f03-42e2-b123-707431218ab6" (UID: "4c0f8247-0f03-42e2-b123-707431218ab6"). InnerVolumeSpecName "kube-api-access-6tdbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.761821 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da5cf300-5392-441c-9cb2-9a7b0268ff1d-kube-api-access-kjxrs" (OuterVolumeSpecName: "kube-api-access-kjxrs") pod "da5cf300-5392-441c-9cb2-9a7b0268ff1d" (UID: "da5cf300-5392-441c-9cb2-9a7b0268ff1d"). InnerVolumeSpecName "kube-api-access-kjxrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.762703 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4c0f8247-0f03-42e2-b123-707431218ab6" (UID: "4c0f8247-0f03-42e2-b123-707431218ab6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.762794 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c0f8247-0f03-42e2-b123-707431218ab6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4c0f8247-0f03-42e2-b123-707431218ab6" (UID: "4c0f8247-0f03-42e2-b123-707431218ab6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.781625 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.855968 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7pd9\" (UniqueName: \"kubernetes.io/projected/a933c882-e11d-46a8-8f67-f77ebc97367c-kube-api-access-h7pd9\") pod \"a933c882-e11d-46a8-8f67-f77ebc97367c\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.856032 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-utilities\") pod \"a933c882-e11d-46a8-8f67-f77ebc97367c\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.856088 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-catalog-content\") pod \"a933c882-e11d-46a8-8f67-f77ebc97367c\" (UID: \"a933c882-e11d-46a8-8f67-f77ebc97367c\") " Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.857057 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-utilities" (OuterVolumeSpecName: "utilities") pod "a933c882-e11d-46a8-8f67-f77ebc97367c" (UID: "a933c882-e11d-46a8-8f67-f77ebc97367c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.859323 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a933c882-e11d-46a8-8f67-f77ebc97367c-kube-api-access-h7pd9" (OuterVolumeSpecName: "kube-api-access-h7pd9") pod "a933c882-e11d-46a8-8f67-f77ebc97367c" (UID: "a933c882-e11d-46a8-8f67-f77ebc97367c"). InnerVolumeSpecName "kube-api-access-h7pd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.860995 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861016 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7pd9\" (UniqueName: \"kubernetes.io/projected/a933c882-e11d-46a8-8f67-f77ebc97367c-kube-api-access-h7pd9\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861029 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c0f8247-0f03-42e2-b123-707431218ab6-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861037 4756 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861046 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjxrs\" (UniqueName: \"kubernetes.io/projected/da5cf300-5392-441c-9cb2-9a7b0268ff1d-kube-api-access-kjxrs\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861053 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861061 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c0f8247-0f03-42e2-b123-707431218ab6-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861068 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5cf300-5392-441c-9cb2-9a7b0268ff1d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861076 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861084 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5cf300-5392-441c-9cb2-9a7b0268ff1d-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.861093 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tdbp\" (UniqueName: \"kubernetes.io/projected/4c0f8247-0f03-42e2-b123-707431218ab6-kube-api-access-6tdbp\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.925802 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a933c882-e11d-46a8-8f67-f77ebc97367c" (UID: "a933c882-e11d-46a8-8f67-f77ebc97367c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:13:32 crc kubenswrapper[4756]: I0203 09:13:32.961850 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a933c882-e11d-46a8-8f67-f77ebc97367c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.107626 4756 generic.go:334] "Generic (PLEG): container finished" podID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerID="fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd" exitCode=0 Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.107694 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zrjwc" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.107668 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrjwc" event={"ID":"a933c882-e11d-46a8-8f67-f77ebc97367c","Type":"ContainerDied","Data":"fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd"} Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.107748 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zrjwc" event={"ID":"a933c882-e11d-46a8-8f67-f77ebc97367c","Type":"ContainerDied","Data":"704345cc223a4f28260edcdcf3d772902276d126d441a7f5e81fddec3c22e9ae"} Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.107773 4756 scope.go:117] "RemoveContainer" containerID="fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.110260 4756 generic.go:334] "Generic (PLEG): container finished" podID="da5cf300-5392-441c-9cb2-9a7b0268ff1d" containerID="6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394" exitCode=0 Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.110332 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" event={"ID":"da5cf300-5392-441c-9cb2-9a7b0268ff1d","Type":"ContainerDied","Data":"6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394"} Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.110355 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" event={"ID":"da5cf300-5392-441c-9cb2-9a7b0268ff1d","Type":"ContainerDied","Data":"e17783777ff6d66013c937662becb067dde567999d3d7bb403273ecfcc374bf7"} Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.110415 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.125788 4756 generic.go:334] "Generic (PLEG): container finished" podID="4c0f8247-0f03-42e2-b123-707431218ab6" containerID="0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00" exitCode=0 Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.125841 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" event={"ID":"4c0f8247-0f03-42e2-b123-707431218ab6","Type":"ContainerDied","Data":"0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00"} Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.125872 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" event={"ID":"4c0f8247-0f03-42e2-b123-707431218ab6","Type":"ContainerDied","Data":"18da30e507446c4fa66f2c775415214ee2ec23061dc8c57d1ecdaa69c22e50c2"} Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.126021 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7784f78584-xhkgl" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.128184 4756 scope.go:117] "RemoveContainer" containerID="86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.136796 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zrjwc"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.146548 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zrjwc"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.165553 4756 scope.go:117] "RemoveContainer" containerID="481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.193631 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.196639 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69c8657f57-q6fpv"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.197896 4756 scope.go:117] "RemoveContainer" containerID="fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.198299 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd\": container with ID starting with fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd not found: ID does not exist" containerID="fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.198323 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd"} err="failed to get container status \"fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd\": rpc error: code = NotFound desc = could not find container \"fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd\": container with ID starting with fc2cca22064f8197b447b16df935ab41c0a20af0f42d6e1cb33563437d47e9cd not found: ID does not exist" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.198355 4756 scope.go:117] "RemoveContainer" containerID="86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.198587 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab\": container with ID starting with 86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab not found: ID does not exist" containerID="86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.198604 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab"} err="failed to get container status \"86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab\": rpc error: code = NotFound desc = could not find container \"86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab\": container with ID starting with 86598ee5fe25e2b804feee62a74bdf03a85c11ff82d109077160d2b94044c4ab not found: ID does not exist" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.198627 4756 scope.go:117] "RemoveContainer" containerID="481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.200511 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2\": container with ID starting with 481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2 not found: ID does not exist" containerID="481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.200541 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2"} err="failed to get container status \"481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2\": rpc error: code = NotFound desc = could not find container \"481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2\": container with ID starting with 481f8c365f4769807a515604c752674d5ccda6a9d7449008ca37ee82f106aed2 not found: ID does not exist" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.200556 4756 scope.go:117] "RemoveContainer" containerID="6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.203563 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7784f78584-xhkgl"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.208527 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7784f78584-xhkgl"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.228261 4756 scope.go:117] "RemoveContainer" containerID="6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.230673 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394\": container with ID starting with 6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394 not found: ID does not exist" containerID="6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.230739 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394"} err="failed to get container status \"6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394\": rpc error: code = NotFound desc = could not find container \"6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394\": container with ID starting with 6b98e536c04d39cdc719d8c799905413e01b9910e3d7b7278a6e4ef0c1824394 not found: ID does not exist" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.230769 4756 scope.go:117] "RemoveContainer" containerID="0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.248725 4756 scope.go:117] "RemoveContainer" containerID="0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.249215 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00\": container with ID starting with 0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00 not found: ID does not exist" containerID="0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.249267 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00"} err="failed to get container status \"0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00\": rpc error: code = NotFound desc = could not find container \"0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00\": container with ID starting with 0f3d7a84df46a4a487721062871808b4222b33ba8f31d98f412df4a056748a00 not found: ID does not exist" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.621095 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c0f8247-0f03-42e2-b123-707431218ab6" path="/var/lib/kubelet/pods/4c0f8247-0f03-42e2-b123-707431218ab6/volumes" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.621820 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" path="/var/lib/kubelet/pods/a933c882-e11d-46a8-8f67-f77ebc97367c/volumes" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.622404 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da5cf300-5392-441c-9cb2-9a7b0268ff1d" path="/var/lib/kubelet/pods/da5cf300-5392-441c-9cb2-9a7b0268ff1d/volumes" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753374 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g"] Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.753618 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c0f8247-0f03-42e2-b123-707431218ab6" containerName="controller-manager" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753629 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c0f8247-0f03-42e2-b123-707431218ab6" containerName="controller-manager" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.753642 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerName="extract-utilities" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753647 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerName="extract-utilities" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.753657 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerName="extract-content" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753663 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerName="extract-content" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.753679 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerName="registry-server" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753685 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerName="registry-server" Feb 03 09:13:33 crc kubenswrapper[4756]: E0203 09:13:33.753699 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da5cf300-5392-441c-9cb2-9a7b0268ff1d" containerName="route-controller-manager" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753705 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="da5cf300-5392-441c-9cb2-9a7b0268ff1d" containerName="route-controller-manager" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753789 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a933c882-e11d-46a8-8f67-f77ebc97367c" containerName="registry-server" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753799 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="da5cf300-5392-441c-9cb2-9a7b0268ff1d" containerName="route-controller-manager" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.753809 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c0f8247-0f03-42e2-b123-707431218ab6" containerName="controller-manager" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.754161 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.756371 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.756405 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.756533 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.757276 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.757271 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.757663 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.761658 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-69cb69855b-5zrkt"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.762494 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.769486 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5356ab43-120c-4078-9736-5e762132c408-serving-cert\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.769532 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-config\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.769561 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2mdb\" (UniqueName: \"kubernetes.io/projected/5356ab43-120c-4078-9736-5e762132c408-kube-api-access-z2mdb\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.769609 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-client-ca\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.771132 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.771403 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.771573 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.773064 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.773250 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.773785 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.775748 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.778948 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69cb69855b-5zrkt"] Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.782571 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.870726 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-config\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.870787 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-config\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.870820 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2mdb\" (UniqueName: \"kubernetes.io/projected/5356ab43-120c-4078-9736-5e762132c408-kube-api-access-z2mdb\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.870862 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k7ws\" (UniqueName: \"kubernetes.io/projected/8dd39528-03d1-4d4a-8d6f-effb336f0063-kube-api-access-9k7ws\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.870886 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-client-ca\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.870928 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-proxy-ca-bundles\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.870962 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-client-ca\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.870995 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd39528-03d1-4d4a-8d6f-effb336f0063-serving-cert\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.871028 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5356ab43-120c-4078-9736-5e762132c408-serving-cert\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.872309 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-client-ca\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.872415 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-config\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.877609 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5356ab43-120c-4078-9736-5e762132c408-serving-cert\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.888226 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2mdb\" (UniqueName: \"kubernetes.io/projected/5356ab43-120c-4078-9736-5e762132c408-kube-api-access-z2mdb\") pod \"route-controller-manager-d598c6797-sng5g\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.972373 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd39528-03d1-4d4a-8d6f-effb336f0063-serving-cert\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.972794 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-config\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.973041 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k7ws\" (UniqueName: \"kubernetes.io/projected/8dd39528-03d1-4d4a-8d6f-effb336f0063-kube-api-access-9k7ws\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.973256 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-client-ca\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.973505 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-proxy-ca-bundles\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.973964 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-config\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.974268 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-client-ca\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.975659 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-proxy-ca-bundles\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.977120 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd39528-03d1-4d4a-8d6f-effb336f0063-serving-cert\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:33 crc kubenswrapper[4756]: I0203 09:13:33.998144 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k7ws\" (UniqueName: \"kubernetes.io/projected/8dd39528-03d1-4d4a-8d6f-effb336f0063-kube-api-access-9k7ws\") pod \"controller-manager-69cb69855b-5zrkt\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:34 crc kubenswrapper[4756]: I0203 09:13:34.076343 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:34 crc kubenswrapper[4756]: I0203 09:13:34.085021 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:34 crc kubenswrapper[4756]: I0203 09:13:34.500091 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69cb69855b-5zrkt"] Feb 03 09:13:34 crc kubenswrapper[4756]: I0203 09:13:34.540574 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g"] Feb 03 09:13:34 crc kubenswrapper[4756]: W0203 09:13:34.551090 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5356ab43_120c_4078_9736_5e762132c408.slice/crio-7a9a1815cf2c9c9fab523491a892de6514e8584b57e0730a16550924754fd6b3 WatchSource:0}: Error finding container 7a9a1815cf2c9c9fab523491a892de6514e8584b57e0730a16550924754fd6b3: Status 404 returned error can't find the container with id 7a9a1815cf2c9c9fab523491a892de6514e8584b57e0730a16550924754fd6b3 Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.145820 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" event={"ID":"5356ab43-120c-4078-9736-5e762132c408","Type":"ContainerStarted","Data":"d62863ba9c94e42de747af520933a865e87f56fd817bb3508f2043c0a41ec83c"} Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.146223 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" event={"ID":"5356ab43-120c-4078-9736-5e762132c408","Type":"ContainerStarted","Data":"7a9a1815cf2c9c9fab523491a892de6514e8584b57e0730a16550924754fd6b3"} Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.147468 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.148885 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" event={"ID":"8dd39528-03d1-4d4a-8d6f-effb336f0063","Type":"ContainerStarted","Data":"99e9ebd7a8deafd26ebf2836fba80952e4f02a164af3849436fe5335f6f91df5"} Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.148911 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" event={"ID":"8dd39528-03d1-4d4a-8d6f-effb336f0063","Type":"ContainerStarted","Data":"6cb7f9345021168f6e3e3090e83c104dfa4f06a170b625a5801205ee8dec9783"} Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.149108 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.155082 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.163683 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" podStartSLOduration=3.163663756 podStartE2EDuration="3.163663756s" podCreationTimestamp="2026-02-03 09:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:35.162314099 +0000 UTC m=+206.312781474" watchObservedRunningTime="2026-02-03 09:13:35.163663756 +0000 UTC m=+206.314131131" Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.180413 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" podStartSLOduration=3.1803935660000002 podStartE2EDuration="3.180393566s" podCreationTimestamp="2026-02-03 09:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:35.178100254 +0000 UTC m=+206.328567629" watchObservedRunningTime="2026-02-03 09:13:35.180393566 +0000 UTC m=+206.330860941" Feb 03 09:13:35 crc kubenswrapper[4756]: I0203 09:13:35.335940 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:42 crc kubenswrapper[4756]: I0203 09:13:42.962844 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" podUID="1daff10d-ff13-4c5a-9433-d33ade5b6486" containerName="oauth-openshift" containerID="cri-o://c0bfc447ce13e025405a965e87ec319c7452d5914c5b5a1a8c9f29c76680feab" gracePeriod=15 Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.196055 4756 generic.go:334] "Generic (PLEG): container finished" podID="1daff10d-ff13-4c5a-9433-d33ade5b6486" containerID="c0bfc447ce13e025405a965e87ec319c7452d5914c5b5a1a8c9f29c76680feab" exitCode=0 Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.196373 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" event={"ID":"1daff10d-ff13-4c5a-9433-d33ade5b6486","Type":"ContainerDied","Data":"c0bfc447ce13e025405a965e87ec319c7452d5914c5b5a1a8c9f29c76680feab"} Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.468230 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501024 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-session\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501079 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-cliconfig\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501104 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-trusted-ca-bundle\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501133 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-policies\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501154 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-provider-selection\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501173 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-login\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501204 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-router-certs\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501226 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-ocp-branding-template\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501251 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-idp-0-file-data\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501287 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-dir\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501302 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-service-ca\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501328 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-error\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501351 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-serving-cert\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.501375 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cmml\" (UniqueName: \"kubernetes.io/projected/1daff10d-ff13-4c5a-9433-d33ade5b6486-kube-api-access-5cmml\") pod \"1daff10d-ff13-4c5a-9433-d33ade5b6486\" (UID: \"1daff10d-ff13-4c5a-9433-d33ade5b6486\") " Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.504204 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.504732 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.504751 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.505239 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.506393 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.506933 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1daff10d-ff13-4c5a-9433-d33ade5b6486-kube-api-access-5cmml" (OuterVolumeSpecName: "kube-api-access-5cmml") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "kube-api-access-5cmml". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.508164 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.508345 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.510504 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.511254 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.514188 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.517363 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.517410 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.517731 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "1daff10d-ff13-4c5a-9433-d33ade5b6486" (UID: "1daff10d-ff13-4c5a-9433-d33ade5b6486"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.565884 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.565959 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.566030 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.566644 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.566718 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872" gracePeriod=600 Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603304 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603336 4756 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603350 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603365 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603377 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603389 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603402 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603414 4756 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1daff10d-ff13-4c5a-9433-d33ade5b6486-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603425 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603437 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603466 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603478 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cmml\" (UniqueName: \"kubernetes.io/projected/1daff10d-ff13-4c5a-9433-d33ade5b6486-kube-api-access-5cmml\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603490 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:43 crc kubenswrapper[4756]: I0203 09:13:43.603503 4756 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1daff10d-ff13-4c5a-9433-d33ade5b6486-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:44 crc kubenswrapper[4756]: I0203 09:13:44.207370 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" event={"ID":"1daff10d-ff13-4c5a-9433-d33ade5b6486","Type":"ContainerDied","Data":"22008310cf66c57e6bd80fcaae6182b187163b39c18451e288da0e760f94159d"} Feb 03 09:13:44 crc kubenswrapper[4756]: I0203 09:13:44.208431 4756 scope.go:117] "RemoveContainer" containerID="c0bfc447ce13e025405a965e87ec319c7452d5914c5b5a1a8c9f29c76680feab" Feb 03 09:13:44 crc kubenswrapper[4756]: I0203 09:13:44.207430 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8jhbb" Feb 03 09:13:44 crc kubenswrapper[4756]: I0203 09:13:44.214395 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872" exitCode=0 Feb 03 09:13:44 crc kubenswrapper[4756]: I0203 09:13:44.214436 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872"} Feb 03 09:13:44 crc kubenswrapper[4756]: I0203 09:13:44.214483 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"a749937f8614549763f69387e47ae3407826bd178fc787c90a020ffe22fb54e1"} Feb 03 09:13:44 crc kubenswrapper[4756]: I0203 09:13:44.246928 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8jhbb"] Feb 03 09:13:44 crc kubenswrapper[4756]: I0203 09:13:44.254239 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8jhbb"] Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.620247 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1daff10d-ff13-4c5a-9433-d33ade5b6486" path="/var/lib/kubelet/pods/1daff10d-ff13-4c5a-9433-d33ade5b6486/volumes" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.758030 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g"] Feb 03 09:13:45 crc kubenswrapper[4756]: E0203 09:13:45.758513 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1daff10d-ff13-4c5a-9433-d33ade5b6486" containerName="oauth-openshift" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.758621 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1daff10d-ff13-4c5a-9433-d33ade5b6486" containerName="oauth-openshift" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.758800 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="1daff10d-ff13-4c5a-9433-d33ade5b6486" containerName="oauth-openshift" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.759280 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.767784 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.768085 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.768805 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.769347 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.769481 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.770768 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.770946 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.772194 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.773256 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.777572 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.777686 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.780807 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.781416 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g"] Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.781485 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.785661 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.791881 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.831243 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.831590 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-audit-policies\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.831715 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.831816 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-login\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.831921 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-session\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832093 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-audit-dir\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832207 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-router-certs\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832252 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832282 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832357 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwfh\" (UniqueName: \"kubernetes.io/projected/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-kube-api-access-djwfh\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832397 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-error\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832443 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-service-ca\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832493 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.832529 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.933971 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934020 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934047 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-audit-policies\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934068 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934085 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-login\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934101 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-session\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934121 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-audit-dir\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934145 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-router-certs\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934164 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934181 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934207 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwfh\" (UniqueName: \"kubernetes.io/projected/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-kube-api-access-djwfh\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934222 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-error\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934243 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-service-ca\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934258 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934716 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-audit-dir\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.934884 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.935740 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.936238 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-service-ca\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.936433 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-audit-policies\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.940139 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.940180 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.940259 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-error\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.940722 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.941533 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-session\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.943710 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-user-template-login\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.943764 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.944223 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-v4-0-config-system-router-certs\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:45 crc kubenswrapper[4756]: I0203 09:13:45.963843 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwfh\" (UniqueName: \"kubernetes.io/projected/21b23890-b2e4-4f84-b3a5-dffcac51c0f7-kube-api-access-djwfh\") pod \"oauth-openshift-7557fdbcd4-l8g9g\" (UID: \"21b23890-b2e4-4f84-b3a5-dffcac51c0f7\") " pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:46 crc kubenswrapper[4756]: I0203 09:13:46.086913 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:46 crc kubenswrapper[4756]: I0203 09:13:46.582541 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g"] Feb 03 09:13:46 crc kubenswrapper[4756]: W0203 09:13:46.589499 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21b23890_b2e4_4f84_b3a5_dffcac51c0f7.slice/crio-f17f851781e2a313be1122967acfde66ee42c56a6da20e4a0ad2b23c456cace0 WatchSource:0}: Error finding container f17f851781e2a313be1122967acfde66ee42c56a6da20e4a0ad2b23c456cace0: Status 404 returned error can't find the container with id f17f851781e2a313be1122967acfde66ee42c56a6da20e4a0ad2b23c456cace0 Feb 03 09:13:47 crc kubenswrapper[4756]: I0203 09:13:47.243285 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" event={"ID":"21b23890-b2e4-4f84-b3a5-dffcac51c0f7","Type":"ContainerStarted","Data":"14387b60411166c39a5902c700e597df7c1c6aa5c0649833504ad9a0cc1ab2db"} Feb 03 09:13:47 crc kubenswrapper[4756]: I0203 09:13:47.243344 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" event={"ID":"21b23890-b2e4-4f84-b3a5-dffcac51c0f7","Type":"ContainerStarted","Data":"f17f851781e2a313be1122967acfde66ee42c56a6da20e4a0ad2b23c456cace0"} Feb 03 09:13:47 crc kubenswrapper[4756]: I0203 09:13:47.243779 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:47 crc kubenswrapper[4756]: I0203 09:13:47.248481 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" Feb 03 09:13:47 crc kubenswrapper[4756]: I0203 09:13:47.264362 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7557fdbcd4-l8g9g" podStartSLOduration=30.264336171 podStartE2EDuration="30.264336171s" podCreationTimestamp="2026-02-03 09:13:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:47.25907731 +0000 UTC m=+218.409544705" watchObservedRunningTime="2026-02-03 09:13:47.264336171 +0000 UTC m=+218.414803566" Feb 03 09:13:51 crc kubenswrapper[4756]: I0203 09:13:51.990026 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-69cb69855b-5zrkt"] Feb 03 09:13:51 crc kubenswrapper[4756]: I0203 09:13:51.990903 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" podUID="8dd39528-03d1-4d4a-8d6f-effb336f0063" containerName="controller-manager" containerID="cri-o://99e9ebd7a8deafd26ebf2836fba80952e4f02a164af3849436fe5335f6f91df5" gracePeriod=30 Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.082665 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g"] Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.082919 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" podUID="5356ab43-120c-4078-9736-5e762132c408" containerName="route-controller-manager" containerID="cri-o://d62863ba9c94e42de747af520933a865e87f56fd817bb3508f2043c0a41ec83c" gracePeriod=30 Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.275269 4756 generic.go:334] "Generic (PLEG): container finished" podID="5356ab43-120c-4078-9736-5e762132c408" containerID="d62863ba9c94e42de747af520933a865e87f56fd817bb3508f2043c0a41ec83c" exitCode=0 Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.275380 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" event={"ID":"5356ab43-120c-4078-9736-5e762132c408","Type":"ContainerDied","Data":"d62863ba9c94e42de747af520933a865e87f56fd817bb3508f2043c0a41ec83c"} Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.279257 4756 generic.go:334] "Generic (PLEG): container finished" podID="8dd39528-03d1-4d4a-8d6f-effb336f0063" containerID="99e9ebd7a8deafd26ebf2836fba80952e4f02a164af3849436fe5335f6f91df5" exitCode=0 Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.279311 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" event={"ID":"8dd39528-03d1-4d4a-8d6f-effb336f0063","Type":"ContainerDied","Data":"99e9ebd7a8deafd26ebf2836fba80952e4f02a164af3849436fe5335f6f91df5"} Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.535264 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.575787 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618371 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-config\") pod \"5356ab43-120c-4078-9736-5e762132c408\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618456 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2mdb\" (UniqueName: \"kubernetes.io/projected/5356ab43-120c-4078-9736-5e762132c408-kube-api-access-z2mdb\") pod \"5356ab43-120c-4078-9736-5e762132c408\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618495 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5356ab43-120c-4078-9736-5e762132c408-serving-cert\") pod \"5356ab43-120c-4078-9736-5e762132c408\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618543 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-config\") pod \"8dd39528-03d1-4d4a-8d6f-effb336f0063\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618585 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-client-ca\") pod \"5356ab43-120c-4078-9736-5e762132c408\" (UID: \"5356ab43-120c-4078-9736-5e762132c408\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618651 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd39528-03d1-4d4a-8d6f-effb336f0063-serving-cert\") pod \"8dd39528-03d1-4d4a-8d6f-effb336f0063\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618682 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-proxy-ca-bundles\") pod \"8dd39528-03d1-4d4a-8d6f-effb336f0063\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618717 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k7ws\" (UniqueName: \"kubernetes.io/projected/8dd39528-03d1-4d4a-8d6f-effb336f0063-kube-api-access-9k7ws\") pod \"8dd39528-03d1-4d4a-8d6f-effb336f0063\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.618738 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-client-ca\") pod \"8dd39528-03d1-4d4a-8d6f-effb336f0063\" (UID: \"8dd39528-03d1-4d4a-8d6f-effb336f0063\") " Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.619455 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-config" (OuterVolumeSpecName: "config") pod "5356ab43-120c-4078-9736-5e762132c408" (UID: "5356ab43-120c-4078-9736-5e762132c408"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.619485 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-client-ca" (OuterVolumeSpecName: "client-ca") pod "8dd39528-03d1-4d4a-8d6f-effb336f0063" (UID: "8dd39528-03d1-4d4a-8d6f-effb336f0063"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.619857 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8dd39528-03d1-4d4a-8d6f-effb336f0063" (UID: "8dd39528-03d1-4d4a-8d6f-effb336f0063"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.619966 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-config" (OuterVolumeSpecName: "config") pod "8dd39528-03d1-4d4a-8d6f-effb336f0063" (UID: "8dd39528-03d1-4d4a-8d6f-effb336f0063"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.620273 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-client-ca" (OuterVolumeSpecName: "client-ca") pod "5356ab43-120c-4078-9736-5e762132c408" (UID: "5356ab43-120c-4078-9736-5e762132c408"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.623752 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dd39528-03d1-4d4a-8d6f-effb336f0063-kube-api-access-9k7ws" (OuterVolumeSpecName: "kube-api-access-9k7ws") pod "8dd39528-03d1-4d4a-8d6f-effb336f0063" (UID: "8dd39528-03d1-4d4a-8d6f-effb336f0063"). InnerVolumeSpecName "kube-api-access-9k7ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.623927 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dd39528-03d1-4d4a-8d6f-effb336f0063-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8dd39528-03d1-4d4a-8d6f-effb336f0063" (UID: "8dd39528-03d1-4d4a-8d6f-effb336f0063"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.624095 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5356ab43-120c-4078-9736-5e762132c408-kube-api-access-z2mdb" (OuterVolumeSpecName: "kube-api-access-z2mdb") pod "5356ab43-120c-4078-9736-5e762132c408" (UID: "5356ab43-120c-4078-9736-5e762132c408"). InnerVolumeSpecName "kube-api-access-z2mdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.629014 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5356ab43-120c-4078-9736-5e762132c408-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5356ab43-120c-4078-9736-5e762132c408" (UID: "5356ab43-120c-4078-9736-5e762132c408"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720676 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720709 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd39528-03d1-4d4a-8d6f-effb336f0063-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720721 4756 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720732 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k7ws\" (UniqueName: \"kubernetes.io/projected/8dd39528-03d1-4d4a-8d6f-effb336f0063-kube-api-access-9k7ws\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720745 4756 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-client-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720755 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5356ab43-120c-4078-9736-5e762132c408-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720765 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2mdb\" (UniqueName: \"kubernetes.io/projected/5356ab43-120c-4078-9736-5e762132c408-kube-api-access-z2mdb\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720776 4756 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5356ab43-120c-4078-9736-5e762132c408-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:52 crc kubenswrapper[4756]: I0203 09:13:52.720786 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd39528-03d1-4d4a-8d6f-effb336f0063-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.315409 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" event={"ID":"8dd39528-03d1-4d4a-8d6f-effb336f0063","Type":"ContainerDied","Data":"6cb7f9345021168f6e3e3090e83c104dfa4f06a170b625a5801205ee8dec9783"} Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.315468 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69cb69855b-5zrkt" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.315482 4756 scope.go:117] "RemoveContainer" containerID="99e9ebd7a8deafd26ebf2836fba80952e4f02a164af3849436fe5335f6f91df5" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.317398 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" event={"ID":"5356ab43-120c-4078-9736-5e762132c408","Type":"ContainerDied","Data":"7a9a1815cf2c9c9fab523491a892de6514e8584b57e0730a16550924754fd6b3"} Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.317523 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.342503 4756 scope.go:117] "RemoveContainer" containerID="d62863ba9c94e42de747af520933a865e87f56fd817bb3508f2043c0a41ec83c" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.354629 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-69cb69855b-5zrkt"] Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.357822 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-69cb69855b-5zrkt"] Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.364485 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g"] Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.370508 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-d598c6797-sng5g"] Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.620651 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5356ab43-120c-4078-9736-5e762132c408" path="/var/lib/kubelet/pods/5356ab43-120c-4078-9736-5e762132c408/volumes" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.621385 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dd39528-03d1-4d4a-8d6f-effb336f0063" path="/var/lib/kubelet/pods/8dd39528-03d1-4d4a-8d6f-effb336f0063/volumes" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.771902 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm"] Feb 03 09:13:53 crc kubenswrapper[4756]: E0203 09:13:53.772232 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd39528-03d1-4d4a-8d6f-effb336f0063" containerName="controller-manager" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.772251 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd39528-03d1-4d4a-8d6f-effb336f0063" containerName="controller-manager" Feb 03 09:13:53 crc kubenswrapper[4756]: E0203 09:13:53.772282 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5356ab43-120c-4078-9736-5e762132c408" containerName="route-controller-manager" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.772294 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5356ab43-120c-4078-9736-5e762132c408" containerName="route-controller-manager" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.772474 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dd39528-03d1-4d4a-8d6f-effb336f0063" containerName="controller-manager" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.772494 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5356ab43-120c-4078-9736-5e762132c408" containerName="route-controller-manager" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.773001 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.776639 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng"] Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.777384 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.777696 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.777945 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.778761 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.778857 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.780214 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm"] Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.780420 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.780558 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.781432 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.781523 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.783363 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng"] Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.783887 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.784158 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.784773 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.785721 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.790757 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834440 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-proxy-ca-bundles\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834553 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxl6f\" (UniqueName: \"kubernetes.io/projected/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-kube-api-access-dxl6f\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834614 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-serving-cert\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834629 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70364f2a-577f-4561-a2c0-d97e99b14cd9-serving-cert\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834661 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-config\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834678 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-client-ca\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834728 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-config\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834769 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mhcv\" (UniqueName: \"kubernetes.io/projected/70364f2a-577f-4561-a2c0-d97e99b14cd9-kube-api-access-2mhcv\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.834794 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-client-ca\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.935846 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-config\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.935903 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-client-ca\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.935955 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-config\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.936002 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mhcv\" (UniqueName: \"kubernetes.io/projected/70364f2a-577f-4561-a2c0-d97e99b14cd9-kube-api-access-2mhcv\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.936038 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-client-ca\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.936071 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-proxy-ca-bundles\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.936111 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxl6f\" (UniqueName: \"kubernetes.io/projected/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-kube-api-access-dxl6f\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.936143 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-serving-cert\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.936164 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70364f2a-577f-4561-a2c0-d97e99b14cd9-serving-cert\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.937231 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-client-ca\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.937271 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-config\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.937467 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-config\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.937756 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-client-ca\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.937845 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70364f2a-577f-4561-a2c0-d97e99b14cd9-proxy-ca-bundles\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.940390 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-serving-cert\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.947799 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70364f2a-577f-4561-a2c0-d97e99b14cd9-serving-cert\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.958167 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxl6f\" (UniqueName: \"kubernetes.io/projected/a05beb2d-a435-46fc-bb24-ae4f4dd22b5e-kube-api-access-dxl6f\") pod \"route-controller-manager-6d646876bd-d4rng\" (UID: \"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e\") " pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:53 crc kubenswrapper[4756]: I0203 09:13:53.958426 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mhcv\" (UniqueName: \"kubernetes.io/projected/70364f2a-577f-4561-a2c0-d97e99b14cd9-kube-api-access-2mhcv\") pod \"controller-manager-6bd8c9884d-gb4sm\" (UID: \"70364f2a-577f-4561-a2c0-d97e99b14cd9\") " pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:54 crc kubenswrapper[4756]: I0203 09:13:54.109700 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:54 crc kubenswrapper[4756]: I0203 09:13:54.120233 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:54 crc kubenswrapper[4756]: I0203 09:13:54.520002 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm"] Feb 03 09:13:54 crc kubenswrapper[4756]: I0203 09:13:54.589199 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng"] Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.351168 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" event={"ID":"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e","Type":"ContainerStarted","Data":"8cc89791aa9e0200a2eb08db3f891275d86ff4a6272e78d7f3cab39a43b7c485"} Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.351531 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.351548 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" event={"ID":"a05beb2d-a435-46fc-bb24-ae4f4dd22b5e","Type":"ContainerStarted","Data":"b07b36e8b629a2210e796d0636e05aabb4e185d5e694770e2dc070819e246416"} Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.370131 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" event={"ID":"70364f2a-577f-4561-a2c0-d97e99b14cd9","Type":"ContainerStarted","Data":"b041a1f461a50b13273c9be62b59fc9161e452ce0a6ef530303e30b31bc05989"} Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.370181 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" event={"ID":"70364f2a-577f-4561-a2c0-d97e99b14cd9","Type":"ContainerStarted","Data":"bd90d4b2e8b042c5bf18658fdf9c7de7bc6c9e8cb54ea4a1f5b1112222828bd4"} Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.370417 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.380567 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.381483 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.397630 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d646876bd-d4rng" podStartSLOduration=3.397612007 podStartE2EDuration="3.397612007s" podCreationTimestamp="2026-02-03 09:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:55.394026301 +0000 UTC m=+226.544493686" watchObservedRunningTime="2026-02-03 09:13:55.397612007 +0000 UTC m=+226.548079382" Feb 03 09:13:55 crc kubenswrapper[4756]: I0203 09:13:55.412508 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bd8c9884d-gb4sm" podStartSLOduration=4.412490577 podStartE2EDuration="4.412490577s" podCreationTimestamp="2026-02-03 09:13:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:13:55.41185066 +0000 UTC m=+226.562318055" watchObservedRunningTime="2026-02-03 09:13:55.412490577 +0000 UTC m=+226.562957952" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.401182 4756 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.402245 4756 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.402533 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692" gracePeriod=15 Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.402649 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2" gracePeriod=15 Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.402636 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1" gracePeriod=15 Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.402668 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5" gracePeriod=15 Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.402693 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02" gracePeriod=15 Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.402827 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403204 4756 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 03 09:14:03 crc kubenswrapper[4756]: E0203 09:14:03.403434 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403458 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 03 09:14:03 crc kubenswrapper[4756]: E0203 09:14:03.403468 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403493 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 03 09:14:03 crc kubenswrapper[4756]: E0203 09:14:03.403503 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403511 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: E0203 09:14:03.403519 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403524 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: E0203 09:14:03.403532 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403538 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 03 09:14:03 crc kubenswrapper[4756]: E0203 09:14:03.403547 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403552 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 03 09:14:03 crc kubenswrapper[4756]: E0203 09:14:03.403559 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403565 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403651 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403883 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403896 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403903 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403911 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.403917 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 03 09:14:03 crc kubenswrapper[4756]: E0203 09:14:03.404004 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.404012 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.404102 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.559180 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.559245 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.559326 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.559383 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.559546 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.559648 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.559701 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.559751 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.661087 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.661192 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.661275 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.661318 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.661927 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.661975 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662052 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662086 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662104 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662176 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662189 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662216 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662329 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662380 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662722 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:03 crc kubenswrapper[4756]: I0203 09:14:03.662725 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.421969 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.423805 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.424427 4756 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2" exitCode=0 Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.424464 4756 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1" exitCode=0 Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.424472 4756 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02" exitCode=0 Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.424486 4756 scope.go:117] "RemoveContainer" containerID="c8640e89d81dcb3858cd0c866d263b03ed0e3f24a5210cdd7f370d66e24b3945" Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.424492 4756 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5" exitCode=2 Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.426176 4756 generic.go:334] "Generic (PLEG): container finished" podID="a6bdec88-abf9-41e5-832e-a3c52065622f" containerID="2ed3bf2bf9f21ab9da027c864b0b2a206898a3da9631f46aff677cce9097317e" exitCode=0 Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.426203 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a6bdec88-abf9-41e5-832e-a3c52065622f","Type":"ContainerDied","Data":"2ed3bf2bf9f21ab9da027c864b0b2a206898a3da9631f46aff677cce9097317e"} Feb 03 09:14:04 crc kubenswrapper[4756]: I0203 09:14:04.426955 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.434372 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.804699 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.805893 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.807150 4756 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.807982 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.852614 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.853358 4756 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.853780 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990291 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6bdec88-abf9-41e5-832e-a3c52065622f-kube-api-access\") pod \"a6bdec88-abf9-41e5-832e-a3c52065622f\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990352 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990372 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990413 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-kubelet-dir\") pod \"a6bdec88-abf9-41e5-832e-a3c52065622f\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990429 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990511 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-var-lock\") pod \"a6bdec88-abf9-41e5-832e-a3c52065622f\" (UID: \"a6bdec88-abf9-41e5-832e-a3c52065622f\") " Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990517 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990547 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a6bdec88-abf9-41e5-832e-a3c52065622f" (UID: "a6bdec88-abf9-41e5-832e-a3c52065622f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990587 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990624 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990672 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-var-lock" (OuterVolumeSpecName: "var-lock") pod "a6bdec88-abf9-41e5-832e-a3c52065622f" (UID: "a6bdec88-abf9-41e5-832e-a3c52065622f"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990796 4756 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990809 4756 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990817 4756 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990826 4756 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.990834 4756 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a6bdec88-abf9-41e5-832e-a3c52065622f-var-lock\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:05 crc kubenswrapper[4756]: I0203 09:14:05.997200 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6bdec88-abf9-41e5-832e-a3c52065622f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a6bdec88-abf9-41e5-832e-a3c52065622f" (UID: "a6bdec88-abf9-41e5-832e-a3c52065622f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.092230 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a6bdec88-abf9-41e5-832e-a3c52065622f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.442815 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.442814 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a6bdec88-abf9-41e5-832e-a3c52065622f","Type":"ContainerDied","Data":"f3b191b0c3242af4350843c36921a19a15814a65da51221ad16cf275fbd08f54"} Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.444059 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3b191b0c3242af4350843c36921a19a15814a65da51221ad16cf275fbd08f54" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.446543 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.447456 4756 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692" exitCode=0 Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.447512 4756 scope.go:117] "RemoveContainer" containerID="6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.447574 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.460952 4756 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.461425 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.467801 4756 scope.go:117] "RemoveContainer" containerID="c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.468920 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.469263 4756 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.483548 4756 scope.go:117] "RemoveContainer" containerID="6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.497980 4756 scope.go:117] "RemoveContainer" containerID="1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.511474 4756 scope.go:117] "RemoveContainer" containerID="c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.529643 4756 scope.go:117] "RemoveContainer" containerID="4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.549022 4756 scope.go:117] "RemoveContainer" containerID="6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2" Feb 03 09:14:06 crc kubenswrapper[4756]: E0203 09:14:06.549568 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\": container with ID starting with 6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2 not found: ID does not exist" containerID="6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.549601 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2"} err="failed to get container status \"6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\": rpc error: code = NotFound desc = could not find container \"6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2\": container with ID starting with 6976dd714380f94aa8f65a96213059f1f6f6ea65633a4f1d2aac82117b2bbbc2 not found: ID does not exist" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.549623 4756 scope.go:117] "RemoveContainer" containerID="c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1" Feb 03 09:14:06 crc kubenswrapper[4756]: E0203 09:14:06.549986 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\": container with ID starting with c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1 not found: ID does not exist" containerID="c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.550030 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1"} err="failed to get container status \"c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\": rpc error: code = NotFound desc = could not find container \"c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1\": container with ID starting with c61e5e5d91b07fb145b5f932cb4b916f6a82a025f7284beeea7fe796748777e1 not found: ID does not exist" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.550045 4756 scope.go:117] "RemoveContainer" containerID="6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02" Feb 03 09:14:06 crc kubenswrapper[4756]: E0203 09:14:06.550300 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\": container with ID starting with 6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02 not found: ID does not exist" containerID="6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.550340 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02"} err="failed to get container status \"6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\": rpc error: code = NotFound desc = could not find container \"6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02\": container with ID starting with 6df7774d2a47fe6f90ffdf6f32f529dbc294d46b2232b992e164fa78da5ffa02 not found: ID does not exist" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.550354 4756 scope.go:117] "RemoveContainer" containerID="1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5" Feb 03 09:14:06 crc kubenswrapper[4756]: E0203 09:14:06.550645 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\": container with ID starting with 1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5 not found: ID does not exist" containerID="1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.550697 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5"} err="failed to get container status \"1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\": rpc error: code = NotFound desc = could not find container \"1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5\": container with ID starting with 1f30fa3a589dd2cc3a23690c400c378721c7f6d67c3d79ea5c6d549e45c856c5 not found: ID does not exist" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.550724 4756 scope.go:117] "RemoveContainer" containerID="c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692" Feb 03 09:14:06 crc kubenswrapper[4756]: E0203 09:14:06.551477 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\": container with ID starting with c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692 not found: ID does not exist" containerID="c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.551505 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692"} err="failed to get container status \"c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\": rpc error: code = NotFound desc = could not find container \"c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692\": container with ID starting with c2414e8cfdb7f7c6c583578a3d6f6ccb1088b53b5d8158b572ca1cd30bd0b692 not found: ID does not exist" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.551554 4756 scope.go:117] "RemoveContainer" containerID="4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4" Feb 03 09:14:06 crc kubenswrapper[4756]: E0203 09:14:06.553531 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\": container with ID starting with 4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4 not found: ID does not exist" containerID="4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4" Feb 03 09:14:06 crc kubenswrapper[4756]: I0203 09:14:06.553573 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4"} err="failed to get container status \"4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\": rpc error: code = NotFound desc = could not find container \"4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4\": container with ID starting with 4feda9e5ced0180c029e4badfb99c3fabfbc729133354b11cfa3ba240aa952b4 not found: ID does not exist" Feb 03 09:14:07 crc kubenswrapper[4756]: I0203 09:14:07.621101 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.438465 4756 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.121:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:08 crc kubenswrapper[4756]: I0203 09:14:08.439282 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.461490 4756 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.121:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1890b1b5ebd9e6c2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-03 09:14:08.460728002 +0000 UTC m=+239.611195377,LastTimestamp:2026-02-03 09:14:08.460728002 +0000 UTC m=+239.611195377,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.758961 4756 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.759684 4756 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.759900 4756 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.760120 4756 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.760386 4756 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:08 crc kubenswrapper[4756]: I0203 09:14:08.760428 4756 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.760740 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="200ms" Feb 03 09:14:08 crc kubenswrapper[4756]: E0203 09:14:08.961561 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="400ms" Feb 03 09:14:09 crc kubenswrapper[4756]: E0203 09:14:09.362617 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="800ms" Feb 03 09:14:09 crc kubenswrapper[4756]: I0203 09:14:09.470885 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253"} Feb 03 09:14:09 crc kubenswrapper[4756]: I0203 09:14:09.470938 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4196b1deb4ac8197d76919301dbe6ba4db981415c59bfa2c8e374c9038e575aa"} Feb 03 09:14:09 crc kubenswrapper[4756]: E0203 09:14:09.471479 4756 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.121:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:09 crc kubenswrapper[4756]: I0203 09:14:09.471625 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:09 crc kubenswrapper[4756]: I0203 09:14:09.617261 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:10 crc kubenswrapper[4756]: E0203 09:14:10.163461 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="1.6s" Feb 03 09:14:11 crc kubenswrapper[4756]: E0203 09:14:11.765101 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="3.2s" Feb 03 09:14:14 crc kubenswrapper[4756]: E0203 09:14:14.966961 4756 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.121:6443: connect: connection refused" interval="6.4s" Feb 03 09:14:15 crc kubenswrapper[4756]: I0203 09:14:15.613793 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:15 crc kubenswrapper[4756]: I0203 09:14:15.615255 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:15 crc kubenswrapper[4756]: I0203 09:14:15.631292 4756 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:15 crc kubenswrapper[4756]: I0203 09:14:15.631667 4756 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:15 crc kubenswrapper[4756]: E0203 09:14:15.632459 4756 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:15 crc kubenswrapper[4756]: I0203 09:14:15.632974 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:15 crc kubenswrapper[4756]: W0203 09:14:15.664712 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-cab94575b3ecfe604cf56ce056e6e4ce051f2ac726fddcec62e679fb33fe6b19 WatchSource:0}: Error finding container cab94575b3ecfe604cf56ce056e6e4ce051f2ac726fddcec62e679fb33fe6b19: Status 404 returned error can't find the container with id cab94575b3ecfe604cf56ce056e6e4ce051f2ac726fddcec62e679fb33fe6b19 Feb 03 09:14:16 crc kubenswrapper[4756]: I0203 09:14:16.511377 4756 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="df6175c7f0df847bc4aa0bae7f8e327e83f1a09a15f67b074bcb35c7c61ff0d0" exitCode=0 Feb 03 09:14:16 crc kubenswrapper[4756]: I0203 09:14:16.511495 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"df6175c7f0df847bc4aa0bae7f8e327e83f1a09a15f67b074bcb35c7c61ff0d0"} Feb 03 09:14:16 crc kubenswrapper[4756]: I0203 09:14:16.512499 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cab94575b3ecfe604cf56ce056e6e4ce051f2ac726fddcec62e679fb33fe6b19"} Feb 03 09:14:16 crc kubenswrapper[4756]: I0203 09:14:16.512969 4756 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:16 crc kubenswrapper[4756]: I0203 09:14:16.513005 4756 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:16 crc kubenswrapper[4756]: E0203 09:14:16.513660 4756 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:16 crc kubenswrapper[4756]: I0203 09:14:16.513686 4756 status_manager.go:851] "Failed to get status for pod" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.121:6443: connect: connection refused" Feb 03 09:14:17 crc kubenswrapper[4756]: I0203 09:14:17.519348 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"99d875ca8ff412426468693be1d4f98233a292378604315e22d6f3ca1650606a"} Feb 03 09:14:17 crc kubenswrapper[4756]: I0203 09:14:17.519706 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7d1914bba4690c8e5852a3f5b3d8e58f5dd6e777138c98553dae004702fb5fc9"} Feb 03 09:14:17 crc kubenswrapper[4756]: I0203 09:14:17.519719 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7597912ff29c0fb2211119c2fba9a226d1e856fd0e0c642bb0164be091e2de40"} Feb 03 09:14:17 crc kubenswrapper[4756]: I0203 09:14:17.519729 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0dc4be59bc27ce99b4dc4c9487274351fc37e74806dfa180bd6cb9229aa97b1f"} Feb 03 09:14:17 crc kubenswrapper[4756]: I0203 09:14:17.524457 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 03 09:14:17 crc kubenswrapper[4756]: I0203 09:14:17.524505 4756 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81" exitCode=1 Feb 03 09:14:17 crc kubenswrapper[4756]: I0203 09:14:17.524533 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81"} Feb 03 09:14:17 crc kubenswrapper[4756]: I0203 09:14:17.525016 4756 scope.go:117] "RemoveContainer" containerID="7778f4995e1d88e0e7df6f42a380e68fc719290c64a67a1fa0d74abcebeeba81" Feb 03 09:14:18 crc kubenswrapper[4756]: I0203 09:14:18.534686 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 03 09:14:18 crc kubenswrapper[4756]: I0203 09:14:18.534764 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"745af1ee2bea03a241a132523320b58f7da941825a21fccdbd0dbbef3e1e15a2"} Feb 03 09:14:18 crc kubenswrapper[4756]: I0203 09:14:18.537740 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d81e8864ae707489bd349c62818770214c6a7d8ebbd18d469ae23dd5297c51b2"} Feb 03 09:14:18 crc kubenswrapper[4756]: I0203 09:14:18.537916 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:18 crc kubenswrapper[4756]: I0203 09:14:18.538093 4756 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:18 crc kubenswrapper[4756]: I0203 09:14:18.538123 4756 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:20 crc kubenswrapper[4756]: I0203 09:14:20.633573 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:20 crc kubenswrapper[4756]: I0203 09:14:20.633976 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:20 crc kubenswrapper[4756]: I0203 09:14:20.637929 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:21 crc kubenswrapper[4756]: I0203 09:14:21.022802 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:14:23 crc kubenswrapper[4756]: I0203 09:14:23.384646 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:14:23 crc kubenswrapper[4756]: I0203 09:14:23.394078 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:14:23 crc kubenswrapper[4756]: I0203 09:14:23.550627 4756 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:24 crc kubenswrapper[4756]: I0203 09:14:24.571794 4756 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:24 crc kubenswrapper[4756]: I0203 09:14:24.571850 4756 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:24 crc kubenswrapper[4756]: I0203 09:14:24.577578 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:24 crc kubenswrapper[4756]: I0203 09:14:24.582237 4756 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a566c748-7f41-4d2c-aa1c-df6d0b4cbac3" Feb 03 09:14:25 crc kubenswrapper[4756]: I0203 09:14:25.576577 4756 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:25 crc kubenswrapper[4756]: I0203 09:14:25.576616 4756 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:29 crc kubenswrapper[4756]: I0203 09:14:29.622894 4756 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a566c748-7f41-4d2c-aa1c-df6d0b4cbac3" Feb 03 09:14:31 crc kubenswrapper[4756]: I0203 09:14:31.027331 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 03 09:14:33 crc kubenswrapper[4756]: I0203 09:14:33.718812 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 03 09:14:33 crc kubenswrapper[4756]: I0203 09:14:33.937547 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 03 09:14:34 crc kubenswrapper[4756]: I0203 09:14:34.126038 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 03 09:14:34 crc kubenswrapper[4756]: I0203 09:14:34.219251 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 03 09:14:34 crc kubenswrapper[4756]: I0203 09:14:34.596905 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 03 09:14:34 crc kubenswrapper[4756]: I0203 09:14:34.894429 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 03 09:14:35 crc kubenswrapper[4756]: I0203 09:14:35.316283 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 03 09:14:35 crc kubenswrapper[4756]: I0203 09:14:35.366314 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 03 09:14:35 crc kubenswrapper[4756]: I0203 09:14:35.612475 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 03 09:14:35 crc kubenswrapper[4756]: I0203 09:14:35.741026 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 03 09:14:35 crc kubenswrapper[4756]: I0203 09:14:35.966723 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 03 09:14:36 crc kubenswrapper[4756]: I0203 09:14:36.050947 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 03 09:14:36 crc kubenswrapper[4756]: I0203 09:14:36.195206 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 03 09:14:36 crc kubenswrapper[4756]: I0203 09:14:36.480144 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 03 09:14:36 crc kubenswrapper[4756]: I0203 09:14:36.622855 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 03 09:14:36 crc kubenswrapper[4756]: I0203 09:14:36.734300 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 03 09:14:36 crc kubenswrapper[4756]: I0203 09:14:36.839731 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 03 09:14:36 crc kubenswrapper[4756]: I0203 09:14:36.909367 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.208643 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.314294 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.359075 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.384033 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.437744 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.455318 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.515238 4756 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.533975 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.558689 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.633226 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.650189 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.651350 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.833912 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 03 09:14:37 crc kubenswrapper[4756]: I0203 09:14:37.919701 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.021779 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.038541 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.048636 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.059257 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.065281 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.090755 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.123515 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.127630 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.436337 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.462978 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.534693 4756 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.655346 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.666765 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.732648 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.754617 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.774759 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.812375 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.826034 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.867959 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.919027 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.923769 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.926830 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 03 09:14:38 crc kubenswrapper[4756]: I0203 09:14:38.966992 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.098735 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.127409 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.132994 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.142874 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.209976 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.238563 4756 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.269885 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.340889 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.379665 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.484958 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.567623 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.580241 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.631401 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.713543 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.716482 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.717583 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.721889 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.749805 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.763228 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.863954 4756 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.868329 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.868593 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.868622 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bltft","openshift-marketplace/redhat-operators-pz2qm","openshift-marketplace/community-operators-v6bml","openshift-marketplace/redhat-marketplace-s64pq","openshift-marketplace/certified-operators-xw49c"] Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.868925 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" podUID="8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" containerName="marketplace-operator" containerID="cri-o://ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1" gracePeriod=30 Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.869755 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xw49c" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerName="registry-server" containerID="cri-o://7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109" gracePeriod=30 Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.870050 4756 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.870080 4756 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c245f03f-c294-456f-915e-a86a4e666fc8" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.870056 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s64pq" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerName="registry-server" containerID="cri-o://9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239" gracePeriod=30 Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.870263 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pz2qm" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="registry-server" containerID="cri-o://8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5" gracePeriod=30 Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.870414 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v6bml" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerName="registry-server" containerID="cri-o://d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55" gracePeriod=30 Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.885177 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.886460 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.913732 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.913713066 podStartE2EDuration="16.913713066s" podCreationTimestamp="2026-02-03 09:14:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:14:39.912672733 +0000 UTC m=+271.063140108" watchObservedRunningTime="2026-02-03 09:14:39.913713066 +0000 UTC m=+271.064180441" Feb 03 09:14:39 crc kubenswrapper[4756]: I0203 09:14:39.958125 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.009755 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.041839 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.054913 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.076502 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.181322 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.181756 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.218664 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5 is running failed: container process not found" containerID="8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5" cmd=["grpc_health_probe","-addr=:50051"] Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.219042 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5 is running failed: container process not found" containerID="8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5" cmd=["grpc_health_probe","-addr=:50051"] Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.219267 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5 is running failed: container process not found" containerID="8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5" cmd=["grpc_health_probe","-addr=:50051"] Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.219301 4756 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-pz2qm" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="registry-server" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.265980 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.289006 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.311824 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.398696 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.407199 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.409399 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.521356 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.529939 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.536628 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.544730 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.548247 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.558123 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65dpn\" (UniqueName: \"kubernetes.io/projected/775639e3-1ecc-4ccd-8207-fb6a557e49b3-kube-api-access-65dpn\") pod \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.558222 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-catalog-content\") pod \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.558251 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-utilities\") pod \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\" (UID: \"775639e3-1ecc-4ccd-8207-fb6a557e49b3\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.559410 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-utilities" (OuterVolumeSpecName: "utilities") pod "775639e3-1ecc-4ccd-8207-fb6a557e49b3" (UID: "775639e3-1ecc-4ccd-8207-fb6a557e49b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.566690 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/775639e3-1ecc-4ccd-8207-fb6a557e49b3-kube-api-access-65dpn" (OuterVolumeSpecName: "kube-api-access-65dpn") pod "775639e3-1ecc-4ccd-8207-fb6a557e49b3" (UID: "775639e3-1ecc-4ccd-8207-fb6a557e49b3"). InnerVolumeSpecName "kube-api-access-65dpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.603734 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.622488 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "775639e3-1ecc-4ccd-8207-fb6a557e49b3" (UID: "775639e3-1ecc-4ccd-8207-fb6a557e49b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.657493 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.658957 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7825\" (UniqueName: \"kubernetes.io/projected/fc312654-25ec-4cf1-b35b-ea49aa157603-kube-api-access-f7825\") pod \"fc312654-25ec-4cf1-b35b-ea49aa157603\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.658986 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-operator-metrics\") pod \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659022 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-catalog-content\") pod \"78ac0913-94cb-4f26-934d-bd005399d428\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659059 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5tfr\" (UniqueName: \"kubernetes.io/projected/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-kube-api-access-t5tfr\") pod \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659083 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg9ls\" (UniqueName: \"kubernetes.io/projected/78ac0913-94cb-4f26-934d-bd005399d428-kube-api-access-sg9ls\") pod \"78ac0913-94cb-4f26-934d-bd005399d428\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659101 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxcht\" (UniqueName: \"kubernetes.io/projected/95a68714-9c47-4457-a37e-90a084d1a2b0-kube-api-access-qxcht\") pod \"95a68714-9c47-4457-a37e-90a084d1a2b0\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659131 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-trusted-ca\") pod \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\" (UID: \"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659153 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-utilities\") pod \"78ac0913-94cb-4f26-934d-bd005399d428\" (UID: \"78ac0913-94cb-4f26-934d-bd005399d428\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659175 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-catalog-content\") pod \"95a68714-9c47-4457-a37e-90a084d1a2b0\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659217 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-utilities\") pod \"95a68714-9c47-4457-a37e-90a084d1a2b0\" (UID: \"95a68714-9c47-4457-a37e-90a084d1a2b0\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659248 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-utilities\") pod \"fc312654-25ec-4cf1-b35b-ea49aa157603\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659288 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-catalog-content\") pod \"fc312654-25ec-4cf1-b35b-ea49aa157603\" (UID: \"fc312654-25ec-4cf1-b35b-ea49aa157603\") " Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659525 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65dpn\" (UniqueName: \"kubernetes.io/projected/775639e3-1ecc-4ccd-8207-fb6a557e49b3-kube-api-access-65dpn\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659541 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659552 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775639e3-1ecc-4ccd-8207-fb6a557e49b3-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.659944 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.660141 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-utilities" (OuterVolumeSpecName: "utilities") pod "fc312654-25ec-4cf1-b35b-ea49aa157603" (UID: "fc312654-25ec-4cf1-b35b-ea49aa157603"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.660189 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-utilities" (OuterVolumeSpecName: "utilities") pod "78ac0913-94cb-4f26-934d-bd005399d428" (UID: "78ac0913-94cb-4f26-934d-bd005399d428"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.660226 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" (UID: "8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.660865 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-utilities" (OuterVolumeSpecName: "utilities") pod "95a68714-9c47-4457-a37e-90a084d1a2b0" (UID: "95a68714-9c47-4457-a37e-90a084d1a2b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.662004 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a68714-9c47-4457-a37e-90a084d1a2b0-kube-api-access-qxcht" (OuterVolumeSpecName: "kube-api-access-qxcht") pod "95a68714-9c47-4457-a37e-90a084d1a2b0" (UID: "95a68714-9c47-4457-a37e-90a084d1a2b0"). InnerVolumeSpecName "kube-api-access-qxcht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.662244 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-kube-api-access-t5tfr" (OuterVolumeSpecName: "kube-api-access-t5tfr") pod "8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" (UID: "8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7"). InnerVolumeSpecName "kube-api-access-t5tfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.662544 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc312654-25ec-4cf1-b35b-ea49aa157603-kube-api-access-f7825" (OuterVolumeSpecName: "kube-api-access-f7825") pod "fc312654-25ec-4cf1-b35b-ea49aa157603" (UID: "fc312654-25ec-4cf1-b35b-ea49aa157603"). InnerVolumeSpecName "kube-api-access-f7825". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.662925 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" (UID: "8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.671943 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78ac0913-94cb-4f26-934d-bd005399d428-kube-api-access-sg9ls" (OuterVolumeSpecName: "kube-api-access-sg9ls") pod "78ac0913-94cb-4f26-934d-bd005399d428" (UID: "78ac0913-94cb-4f26-934d-bd005399d428"). InnerVolumeSpecName "kube-api-access-sg9ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.683915 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc312654-25ec-4cf1-b35b-ea49aa157603" (UID: "fc312654-25ec-4cf1-b35b-ea49aa157603"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.688857 4756 generic.go:334] "Generic (PLEG): container finished" podID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerID="9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239" exitCode=0 Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.688944 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s64pq" event={"ID":"fc312654-25ec-4cf1-b35b-ea49aa157603","Type":"ContainerDied","Data":"9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.688978 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s64pq" event={"ID":"fc312654-25ec-4cf1-b35b-ea49aa157603","Type":"ContainerDied","Data":"c4301705640587639a286214d9eead21b0ab5aaf2c99f2705a0da284980fd69c"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.689006 4756 scope.go:117] "RemoveContainer" containerID="9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.689145 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s64pq" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.692963 4756 generic.go:334] "Generic (PLEG): container finished" podID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerID="7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109" exitCode=0 Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.693025 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw49c" event={"ID":"95a68714-9c47-4457-a37e-90a084d1a2b0","Type":"ContainerDied","Data":"7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.693055 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xw49c" event={"ID":"95a68714-9c47-4457-a37e-90a084d1a2b0","Type":"ContainerDied","Data":"6ddab397bc19c427cbf60f38f2590c655a17376aaab584a70c17c5f2327321bf"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.693262 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xw49c" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.695220 4756 generic.go:334] "Generic (PLEG): container finished" podID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerID="d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55" exitCode=0 Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.695341 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6bml" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.695988 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6bml" event={"ID":"775639e3-1ecc-4ccd-8207-fb6a557e49b3","Type":"ContainerDied","Data":"d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.696030 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6bml" event={"ID":"775639e3-1ecc-4ccd-8207-fb6a557e49b3","Type":"ContainerDied","Data":"072c6adc1547fdbf15aa241e57909345bec616cfa902b47d60cf7655380defa0"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.698533 4756 generic.go:334] "Generic (PLEG): container finished" podID="8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" containerID="ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1" exitCode=0 Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.698590 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" event={"ID":"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7","Type":"ContainerDied","Data":"ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.698610 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" event={"ID":"8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7","Type":"ContainerDied","Data":"03765366fd902afb2e3296b98dcda116ed0ca540ee2a743629c001158c87cb39"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.698651 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bltft" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.701356 4756 generic.go:334] "Generic (PLEG): container finished" podID="78ac0913-94cb-4f26-934d-bd005399d428" containerID="8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5" exitCode=0 Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.701557 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz2qm" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.701583 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz2qm" event={"ID":"78ac0913-94cb-4f26-934d-bd005399d428","Type":"ContainerDied","Data":"8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.701600 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz2qm" event={"ID":"78ac0913-94cb-4f26-934d-bd005399d428","Type":"ContainerDied","Data":"367617e4713272129bdb888c9eed5b28230d7259bf332e9b8e67c0cc16f99e60"} Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.706059 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.709023 4756 scope.go:117] "RemoveContainer" containerID="ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.722511 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s64pq"] Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.727505 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s64pq"] Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.733341 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95a68714-9c47-4457-a37e-90a084d1a2b0" (UID: "95a68714-9c47-4457-a37e-90a084d1a2b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.733697 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bltft"] Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.736620 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bltft"] Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.741210 4756 scope.go:117] "RemoveContainer" containerID="76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.754633 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6bml"] Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.756652 4756 scope.go:117] "RemoveContainer" containerID="9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.757154 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239\": container with ID starting with 9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239 not found: ID does not exist" containerID="9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.757193 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239"} err="failed to get container status \"9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239\": rpc error: code = NotFound desc = could not find container \"9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239\": container with ID starting with 9205e5ba217f1a243375e3ffeed1406f96972883f09e2ca98e93b8fc8cd78239 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.757218 4756 scope.go:117] "RemoveContainer" containerID="ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.757587 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488\": container with ID starting with ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488 not found: ID does not exist" containerID="ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.757618 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488"} err="failed to get container status \"ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488\": rpc error: code = NotFound desc = could not find container \"ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488\": container with ID starting with ffa6b84e034362dc3e5c00f063fd978b1c8959f0218156d08ea221b52a321488 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.757639 4756 scope.go:117] "RemoveContainer" containerID="76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.757833 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb\": container with ID starting with 76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb not found: ID does not exist" containerID="76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.757934 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb"} err="failed to get container status \"76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb\": rpc error: code = NotFound desc = could not find container \"76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb\": container with ID starting with 76e70cdbd681bed71caaa524619cbe107c8be6b82be25b3d8f32304cb9e24afb not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.758042 4756 scope.go:117] "RemoveContainer" containerID="7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.757984 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v6bml"] Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760340 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760359 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc312654-25ec-4cf1-b35b-ea49aa157603-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760390 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7825\" (UniqueName: \"kubernetes.io/projected/fc312654-25ec-4cf1-b35b-ea49aa157603-kube-api-access-f7825\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760401 4756 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760410 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5tfr\" (UniqueName: \"kubernetes.io/projected/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-kube-api-access-t5tfr\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760418 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxcht\" (UniqueName: \"kubernetes.io/projected/95a68714-9c47-4457-a37e-90a084d1a2b0-kube-api-access-qxcht\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760426 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg9ls\" (UniqueName: \"kubernetes.io/projected/78ac0913-94cb-4f26-934d-bd005399d428-kube-api-access-sg9ls\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760435 4756 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760462 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760470 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.760480 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a68714-9c47-4457-a37e-90a084d1a2b0-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.769301 4756 scope.go:117] "RemoveContainer" containerID="57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.788574 4756 scope.go:117] "RemoveContainer" containerID="6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.796742 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.801712 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78ac0913-94cb-4f26-934d-bd005399d428" (UID: "78ac0913-94cb-4f26-934d-bd005399d428"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.802344 4756 scope.go:117] "RemoveContainer" containerID="7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.802818 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109\": container with ID starting with 7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109 not found: ID does not exist" containerID="7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.802847 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109"} err="failed to get container status \"7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109\": rpc error: code = NotFound desc = could not find container \"7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109\": container with ID starting with 7fe503fba8497372fb3683f12d6311c5566c7e92e7a35245660ef407cd9d3109 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.802867 4756 scope.go:117] "RemoveContainer" containerID="57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.803218 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e\": container with ID starting with 57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e not found: ID does not exist" containerID="57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.803262 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e"} err="failed to get container status \"57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e\": rpc error: code = NotFound desc = could not find container \"57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e\": container with ID starting with 57917382d3318e405ef97a13b2758d9656bc5730b0aff95546727b33069a6b0e not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.803281 4756 scope.go:117] "RemoveContainer" containerID="6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.803756 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433\": container with ID starting with 6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433 not found: ID does not exist" containerID="6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.803781 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433"} err="failed to get container status \"6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433\": rpc error: code = NotFound desc = could not find container \"6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433\": container with ID starting with 6756024cac0e3b4aeee8905a1e9c404cf67627a7e296103bee5c0dedd54d5433 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.803817 4756 scope.go:117] "RemoveContainer" containerID="d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.808612 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.816585 4756 scope.go:117] "RemoveContainer" containerID="9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.833030 4756 scope.go:117] "RemoveContainer" containerID="c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.844113 4756 scope.go:117] "RemoveContainer" containerID="d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.844715 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55\": container with ID starting with d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55 not found: ID does not exist" containerID="d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.844752 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55"} err="failed to get container status \"d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55\": rpc error: code = NotFound desc = could not find container \"d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55\": container with ID starting with d27ce8c95e8c6487222feefdc5221a5eac84c656be14db1fe944dac4855c1d55 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.844779 4756 scope.go:117] "RemoveContainer" containerID="9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.845928 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518\": container with ID starting with 9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518 not found: ID does not exist" containerID="9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.845966 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518"} err="failed to get container status \"9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518\": rpc error: code = NotFound desc = could not find container \"9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518\": container with ID starting with 9d98568b1b4a14cb6057d9b92446e2205cbcdd5020228c8a35434a484e4c2518 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.846055 4756 scope.go:117] "RemoveContainer" containerID="c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.846854 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409\": container with ID starting with c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409 not found: ID does not exist" containerID="c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.847106 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409"} err="failed to get container status \"c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409\": rpc error: code = NotFound desc = could not find container \"c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409\": container with ID starting with c134e2990b1cb0cb8d5d26cf38c54c64cff9f1be8a8edb249e6f3ec844737409 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.847166 4756 scope.go:117] "RemoveContainer" containerID="ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.859621 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.859985 4756 scope.go:117] "RemoveContainer" containerID="ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.860984 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78ac0913-94cb-4f26-934d-bd005399d428-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.861248 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1\": container with ID starting with ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1 not found: ID does not exist" containerID="ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.861277 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1"} err="failed to get container status \"ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1\": rpc error: code = NotFound desc = could not find container \"ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1\": container with ID starting with ac2123f1a72b658433a4bb0ae4450923dcba6473ed8fa0bdc75ff69b04cadfe1 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.861303 4756 scope.go:117] "RemoveContainer" containerID="8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.877762 4756 scope.go:117] "RemoveContainer" containerID="eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.895026 4756 scope.go:117] "RemoveContainer" containerID="bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.911424 4756 scope.go:117] "RemoveContainer" containerID="8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.911864 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5\": container with ID starting with 8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5 not found: ID does not exist" containerID="8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.911918 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5"} err="failed to get container status \"8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5\": rpc error: code = NotFound desc = could not find container \"8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5\": container with ID starting with 8f09e2c34a6c6d479479f3ae3e08e1fa5bbd856a4a1448e7fda558cb504c9de5 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.911949 4756 scope.go:117] "RemoveContainer" containerID="eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.912536 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85\": container with ID starting with eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85 not found: ID does not exist" containerID="eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.912587 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85"} err="failed to get container status \"eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85\": rpc error: code = NotFound desc = could not find container \"eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85\": container with ID starting with eedb361a05d2fdb5da7976b5d64314e9ef0475db3f5a0befddffbedb41c3af85 not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.912619 4756 scope.go:117] "RemoveContainer" containerID="bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d" Feb 03 09:14:40 crc kubenswrapper[4756]: E0203 09:14:40.912868 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d\": container with ID starting with bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d not found: ID does not exist" containerID="bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.912888 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d"} err="failed to get container status \"bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d\": rpc error: code = NotFound desc = could not find container \"bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d\": container with ID starting with bbc0e565583f7b05afac500c07c0534d8698ea6529d74c11b8d7ddc642811d2d not found: ID does not exist" Feb 03 09:14:40 crc kubenswrapper[4756]: I0203 09:14:40.998206 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.017604 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.022249 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xw49c"] Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.032865 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xw49c"] Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.041131 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pz2qm"] Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.044282 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pz2qm"] Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.145389 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.153633 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.192624 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.287788 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.295824 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.317634 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.341372 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.376573 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.504948 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.547147 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.548690 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.569963 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.621354 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" path="/var/lib/kubelet/pods/775639e3-1ecc-4ccd-8207-fb6a557e49b3/volumes" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.622851 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78ac0913-94cb-4f26-934d-bd005399d428" path="/var/lib/kubelet/pods/78ac0913-94cb-4f26-934d-bd005399d428/volumes" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.626398 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" path="/var/lib/kubelet/pods/8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7/volumes" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.628880 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" path="/var/lib/kubelet/pods/95a68714-9c47-4457-a37e-90a084d1a2b0/volumes" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.636366 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" path="/var/lib/kubelet/pods/fc312654-25ec-4cf1-b35b-ea49aa157603/volumes" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.645191 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.741987 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.790841 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.792422 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.803307 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 03 09:14:41 crc kubenswrapper[4756]: I0203 09:14:41.996391 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.059375 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.109023 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.134702 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.261771 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.353599 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.430426 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.593758 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.640896 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.722746 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.737895 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.743320 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.757201 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.882959 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.898280 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.920203 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 03 09:14:42 crc kubenswrapper[4756]: I0203 09:14:42.953120 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.056886 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.082894 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.087429 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.091533 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.097461 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.121031 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.168734 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.233426 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.338551 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.381589 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.606289 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.627076 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.639812 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.642840 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.667988 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.684114 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.748100 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.893415 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 03 09:14:43 crc kubenswrapper[4756]: I0203 09:14:43.956008 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.020932 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.050246 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.104228 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.135757 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.181047 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.190814 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.268394 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.278655 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.344490 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.587667 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.598941 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.651966 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.688384 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.890339 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 03 09:14:44 crc kubenswrapper[4756]: I0203 09:14:44.995204 4756 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.038070 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.052877 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.128606 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.158689 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.245406 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.252781 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.266227 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.364334 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.407923 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.455107 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.488930 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.594844 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.604105 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.659202 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.762760 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.772644 4756 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.772895 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253" gracePeriod=5 Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813525 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xztlg"] Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813749 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerName="extract-content" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813765 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerName="extract-content" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813780 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" containerName="installer" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813789 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" containerName="installer" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813800 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813810 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813823 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerName="extract-content" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813831 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerName="extract-content" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813841 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813848 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813858 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerName="extract-utilities" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813866 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerName="extract-utilities" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813878 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" containerName="marketplace-operator" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813886 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" containerName="marketplace-operator" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813897 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="extract-content" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813904 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="extract-content" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813914 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="extract-utilities" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813921 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="extract-utilities" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813931 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813937 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813949 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813956 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813964 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerName="extract-content" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813971 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerName="extract-content" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813981 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerName="extract-utilities" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.813988 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerName="extract-utilities" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.813997 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerName="extract-utilities" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814364 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerName="extract-utilities" Feb 03 09:14:45 crc kubenswrapper[4756]: E0203 09:14:45.814384 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814394 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814519 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="775639e3-1ecc-4ccd-8207-fb6a557e49b3" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814533 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6bdec88-abf9-41e5-832e-a3c52065622f" containerName="installer" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814545 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f92ee5e-7d6d-4d0c-8f83-5181c2d9fba7" containerName="marketplace-operator" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814558 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc312654-25ec-4cf1-b35b-ea49aa157603" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814567 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="78ac0913-94cb-4f26-934d-bd005399d428" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814577 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a68714-9c47-4457-a37e-90a084d1a2b0" containerName="registry-server" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.814585 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.815062 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.816548 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/29956f45-fbd4-4630-82fe-8253dc497ce6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.816613 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29956f45-fbd4-4630-82fe-8253dc497ce6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.816676 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdlls\" (UniqueName: \"kubernetes.io/projected/29956f45-fbd4-4630-82fe-8253dc497ce6-kube-api-access-tdlls\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.817194 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.817371 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.817554 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.819794 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.821807 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xztlg"] Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.822553 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.906258 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.917252 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdlls\" (UniqueName: \"kubernetes.io/projected/29956f45-fbd4-4630-82fe-8253dc497ce6-kube-api-access-tdlls\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.917335 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/29956f45-fbd4-4630-82fe-8253dc497ce6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.917360 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29956f45-fbd4-4630-82fe-8253dc497ce6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.918857 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29956f45-fbd4-4630-82fe-8253dc497ce6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.924128 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/29956f45-fbd4-4630-82fe-8253dc497ce6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:45 crc kubenswrapper[4756]: I0203 09:14:45.939895 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdlls\" (UniqueName: \"kubernetes.io/projected/29956f45-fbd4-4630-82fe-8253dc497ce6-kube-api-access-tdlls\") pod \"marketplace-operator-79b997595-xztlg\" (UID: \"29956f45-fbd4-4630-82fe-8253dc497ce6\") " pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.011320 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.026556 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.128868 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.147090 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.163348 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.207427 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.274403 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.274627 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.316311 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.583638 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.590195 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xztlg"] Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.593893 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.596074 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.610988 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.612485 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.683125 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.737282 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" event={"ID":"29956f45-fbd4-4630-82fe-8253dc497ce6","Type":"ContainerStarted","Data":"d1553e1c0c38033be36e14deff8f12bd8dac3b7450da5baa18bd721a74c8706b"} Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.737334 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" event={"ID":"29956f45-fbd4-4630-82fe-8253dc497ce6","Type":"ContainerStarted","Data":"e12ce2db4b165f69e77644c620d5e470805e1ffcd94939e6aa5aa0a986a66918"} Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.737682 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.739039 4756 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xztlg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.66:8080/healthz\": dial tcp 10.217.0.66:8080: connect: connection refused" start-of-body= Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.739085 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" podUID="29956f45-fbd4-4630-82fe-8253dc497ce6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.66:8080/healthz\": dial tcp 10.217.0.66:8080: connect: connection refused" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.743043 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.755114 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.758606 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" podStartSLOduration=6.758585871 podStartE2EDuration="6.758585871s" podCreationTimestamp="2026-02-03 09:14:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:14:46.755070817 +0000 UTC m=+277.905538192" watchObservedRunningTime="2026-02-03 09:14:46.758585871 +0000 UTC m=+277.909053246" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.858073 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.922378 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 03 09:14:46 crc kubenswrapper[4756]: I0203 09:14:46.931045 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.004883 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.017632 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.071029 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.110186 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.163738 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.255851 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.263350 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.287171 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.534906 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.632517 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.667554 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.670300 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.745148 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xztlg" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.749031 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.777478 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.951279 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 03 09:14:47 crc kubenswrapper[4756]: I0203 09:14:47.999641 4756 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.040541 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.111406 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.142784 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.283292 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.400219 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.476018 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.543806 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.586807 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.596717 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.719588 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.765789 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.818206 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 03 09:14:48 crc kubenswrapper[4756]: I0203 09:14:48.993255 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 03 09:14:49 crc kubenswrapper[4756]: I0203 09:14:49.093838 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 03 09:14:49 crc kubenswrapper[4756]: I0203 09:14:49.244032 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 03 09:14:49 crc kubenswrapper[4756]: I0203 09:14:49.357311 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 03 09:14:49 crc kubenswrapper[4756]: I0203 09:14:49.419297 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 03 09:14:49 crc kubenswrapper[4756]: I0203 09:14:49.461011 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 03 09:14:49 crc kubenswrapper[4756]: I0203 09:14:49.637370 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 03 09:14:49 crc kubenswrapper[4756]: I0203 09:14:49.924553 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 03 09:14:50 crc kubenswrapper[4756]: I0203 09:14:50.222355 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 03 09:14:50 crc kubenswrapper[4756]: I0203 09:14:50.243689 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 03 09:14:50 crc kubenswrapper[4756]: I0203 09:14:50.370151 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 03 09:14:50 crc kubenswrapper[4756]: I0203 09:14:50.803051 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.044327 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.335981 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.336058 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.398930 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.398995 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399077 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399128 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399146 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399175 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399233 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399246 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399274 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399644 4756 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399669 4756 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399678 4756 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.399685 4756 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.407573 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.500550 4756 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.558731 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.567887 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.619588 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.776933 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.776994 4756 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253" exitCode=137 Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.777067 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.777067 4756 scope.go:117] "RemoveContainer" containerID="02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.799259 4756 scope.go:117] "RemoveContainer" containerID="02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253" Feb 03 09:14:51 crc kubenswrapper[4756]: E0203 09:14:51.799720 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253\": container with ID starting with 02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253 not found: ID does not exist" containerID="02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.799760 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253"} err="failed to get container status \"02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253\": rpc error: code = NotFound desc = could not find container \"02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253\": container with ID starting with 02ec54440f2db6a2da09f49b3bf3080d8d4124325de7005cde82f5265ecb4253 not found: ID does not exist" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.800783 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 03 09:14:51 crc kubenswrapper[4756]: I0203 09:14:51.812133 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 03 09:14:52 crc kubenswrapper[4756]: I0203 09:14:52.440845 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.167737 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn"] Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.169114 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.171932 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.172116 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.177426 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn"] Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.202003 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjmx5\" (UniqueName: \"kubernetes.io/projected/ac16e184-042c-43c9-8848-b549fb5e44a2-kube-api-access-hjmx5\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.202759 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac16e184-042c-43c9-8848-b549fb5e44a2-config-volume\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.202842 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac16e184-042c-43c9-8848-b549fb5e44a2-secret-volume\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.303316 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjmx5\" (UniqueName: \"kubernetes.io/projected/ac16e184-042c-43c9-8848-b549fb5e44a2-kube-api-access-hjmx5\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.303359 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac16e184-042c-43c9-8848-b549fb5e44a2-config-volume\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.303394 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac16e184-042c-43c9-8848-b549fb5e44a2-secret-volume\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.304259 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac16e184-042c-43c9-8848-b549fb5e44a2-config-volume\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.309369 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac16e184-042c-43c9-8848-b549fb5e44a2-secret-volume\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.321763 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjmx5\" (UniqueName: \"kubernetes.io/projected/ac16e184-042c-43c9-8848-b549fb5e44a2-kube-api-access-hjmx5\") pod \"collect-profiles-29501835-p9bwn\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.497710 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:00 crc kubenswrapper[4756]: I0203 09:15:00.932105 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn"] Feb 03 09:15:01 crc kubenswrapper[4756]: I0203 09:15:01.924050 4756 generic.go:334] "Generic (PLEG): container finished" podID="ac16e184-042c-43c9-8848-b549fb5e44a2" containerID="bbdc33cc9fa87a5c29ad593bdcd5b4c97b1b35a614bfab26b39ccdb4ca22ab15" exitCode=0 Feb 03 09:15:01 crc kubenswrapper[4756]: I0203 09:15:01.924095 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" event={"ID":"ac16e184-042c-43c9-8848-b549fb5e44a2","Type":"ContainerDied","Data":"bbdc33cc9fa87a5c29ad593bdcd5b4c97b1b35a614bfab26b39ccdb4ca22ab15"} Feb 03 09:15:01 crc kubenswrapper[4756]: I0203 09:15:01.924120 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" event={"ID":"ac16e184-042c-43c9-8848-b549fb5e44a2","Type":"ContainerStarted","Data":"a556c9fd5109dd8368347e06c74ae6629240f1343bb533bf249e7d980853a008"} Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.178087 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.236595 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac16e184-042c-43c9-8848-b549fb5e44a2-config-volume\") pod \"ac16e184-042c-43c9-8848-b549fb5e44a2\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.236701 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac16e184-042c-43c9-8848-b549fb5e44a2-secret-volume\") pod \"ac16e184-042c-43c9-8848-b549fb5e44a2\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.236720 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjmx5\" (UniqueName: \"kubernetes.io/projected/ac16e184-042c-43c9-8848-b549fb5e44a2-kube-api-access-hjmx5\") pod \"ac16e184-042c-43c9-8848-b549fb5e44a2\" (UID: \"ac16e184-042c-43c9-8848-b549fb5e44a2\") " Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.237281 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac16e184-042c-43c9-8848-b549fb5e44a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "ac16e184-042c-43c9-8848-b549fb5e44a2" (UID: "ac16e184-042c-43c9-8848-b549fb5e44a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.237563 4756 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac16e184-042c-43c9-8848-b549fb5e44a2-config-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.241110 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac16e184-042c-43c9-8848-b549fb5e44a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ac16e184-042c-43c9-8848-b549fb5e44a2" (UID: "ac16e184-042c-43c9-8848-b549fb5e44a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.241113 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac16e184-042c-43c9-8848-b549fb5e44a2-kube-api-access-hjmx5" (OuterVolumeSpecName: "kube-api-access-hjmx5") pod "ac16e184-042c-43c9-8848-b549fb5e44a2" (UID: "ac16e184-042c-43c9-8848-b549fb5e44a2"). InnerVolumeSpecName "kube-api-access-hjmx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.338524 4756 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac16e184-042c-43c9-8848-b549fb5e44a2-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.338565 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjmx5\" (UniqueName: \"kubernetes.io/projected/ac16e184-042c-43c9-8848-b549fb5e44a2-kube-api-access-hjmx5\") on node \"crc\" DevicePath \"\"" Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.937097 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" event={"ID":"ac16e184-042c-43c9-8848-b549fb5e44a2","Type":"ContainerDied","Data":"a556c9fd5109dd8368347e06c74ae6629240f1343bb533bf249e7d980853a008"} Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.937153 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a556c9fd5109dd8368347e06c74ae6629240f1343bb533bf249e7d980853a008" Feb 03 09:15:03 crc kubenswrapper[4756]: I0203 09:15:03.937176 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn" Feb 03 09:15:09 crc kubenswrapper[4756]: I0203 09:15:09.383795 4756 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 03 09:15:43 crc kubenswrapper[4756]: I0203 09:15:43.566614 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:15:43 crc kubenswrapper[4756]: I0203 09:15:43.567570 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.588888 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dvqhq"] Feb 03 09:15:57 crc kubenswrapper[4756]: E0203 09:15:57.589367 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac16e184-042c-43c9-8848-b549fb5e44a2" containerName="collect-profiles" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.589378 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac16e184-042c-43c9-8848-b549fb5e44a2" containerName="collect-profiles" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.589497 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac16e184-042c-43c9-8848-b549fb5e44a2" containerName="collect-profiles" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.590215 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.596393 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.597706 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvqhq"] Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.697041 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad66a22-bd41-40e4-b88f-d222ff093a15-utilities\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.697116 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvgqn\" (UniqueName: \"kubernetes.io/projected/4ad66a22-bd41-40e4-b88f-d222ff093a15-kube-api-access-mvgqn\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.697216 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad66a22-bd41-40e4-b88f-d222ff093a15-catalog-content\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.783239 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h5999"] Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.784202 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.786165 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.795792 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5999"] Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.797964 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvgqn\" (UniqueName: \"kubernetes.io/projected/4ad66a22-bd41-40e4-b88f-d222ff093a15-kube-api-access-mvgqn\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.798027 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad66a22-bd41-40e4-b88f-d222ff093a15-catalog-content\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.798069 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad66a22-bd41-40e4-b88f-d222ff093a15-utilities\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.798543 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ad66a22-bd41-40e4-b88f-d222ff093a15-utilities\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.798759 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ad66a22-bd41-40e4-b88f-d222ff093a15-catalog-content\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.830737 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvgqn\" (UniqueName: \"kubernetes.io/projected/4ad66a22-bd41-40e4-b88f-d222ff093a15-kube-api-access-mvgqn\") pod \"redhat-operators-dvqhq\" (UID: \"4ad66a22-bd41-40e4-b88f-d222ff093a15\") " pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.899588 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3021f6bf-d879-49e1-8051-9f599be91475-utilities\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.899640 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbccj\" (UniqueName: \"kubernetes.io/projected/3021f6bf-d879-49e1-8051-9f599be91475-kube-api-access-wbccj\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.899802 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3021f6bf-d879-49e1-8051-9f599be91475-catalog-content\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.905532 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.984708 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r8qjq"] Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.985433 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:57 crc kubenswrapper[4756]: I0203 09:15:57.998463 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r8qjq"] Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.000803 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3021f6bf-d879-49e1-8051-9f599be91475-utilities\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.000843 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbccj\" (UniqueName: \"kubernetes.io/projected/3021f6bf-d879-49e1-8051-9f599be91475-kube-api-access-wbccj\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.000868 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3021f6bf-d879-49e1-8051-9f599be91475-catalog-content\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.001296 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3021f6bf-d879-49e1-8051-9f599be91475-catalog-content\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.001601 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3021f6bf-d879-49e1-8051-9f599be91475-utilities\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.031370 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbccj\" (UniqueName: \"kubernetes.io/projected/3021f6bf-d879-49e1-8051-9f599be91475-kube-api-access-wbccj\") pod \"redhat-marketplace-h5999\" (UID: \"3021f6bf-d879-49e1-8051-9f599be91475\") " pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.100201 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.103664 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.103717 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-registry-tls\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.103735 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-registry-certificates\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.103764 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzspt\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-kube-api-access-rzspt\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.103796 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-trusted-ca\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.103830 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-bound-sa-token\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.103862 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.103910 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.133748 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.205421 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-registry-tls\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.205520 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-registry-certificates\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.205545 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzspt\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-kube-api-access-rzspt\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.205576 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-trusted-ca\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.205603 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-bound-sa-token\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.205645 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.205672 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.206063 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.206708 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-registry-certificates\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.207899 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-trusted-ca\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.209690 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.210263 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-registry-tls\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.223489 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzspt\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-kube-api-access-rzspt\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.223979 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3c0e7700-9f5b-4ece-8e9d-234c14d72a76-bound-sa-token\") pod \"image-registry-66df7c8f76-r8qjq\" (UID: \"3c0e7700-9f5b-4ece-8e9d-234c14d72a76\") " pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.271216 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5999"] Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.303503 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.339948 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvqhq"] Feb 03 09:15:58 crc kubenswrapper[4756]: W0203 09:15:58.343767 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad66a22_bd41_40e4_b88f_d222ff093a15.slice/crio-f3b44280748c4f835e9f681e0f69b4e877f12e3cd1b1ea52b6881555ebec60fa WatchSource:0}: Error finding container f3b44280748c4f835e9f681e0f69b4e877f12e3cd1b1ea52b6881555ebec60fa: Status 404 returned error can't find the container with id f3b44280748c4f835e9f681e0f69b4e877f12e3cd1b1ea52b6881555ebec60fa Feb 03 09:15:58 crc kubenswrapper[4756]: I0203 09:15:58.487389 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r8qjq"] Feb 03 09:15:58 crc kubenswrapper[4756]: W0203 09:15:58.547587 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c0e7700_9f5b_4ece_8e9d_234c14d72a76.slice/crio-3e0e824189bbf24fd072a12cbfb36d79369cf99f60f91b94bfe893e592c3442c WatchSource:0}: Error finding container 3e0e824189bbf24fd072a12cbfb36d79369cf99f60f91b94bfe893e592c3442c: Status 404 returned error can't find the container with id 3e0e824189bbf24fd072a12cbfb36d79369cf99f60f91b94bfe893e592c3442c Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.232145 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" event={"ID":"3c0e7700-9f5b-4ece-8e9d-234c14d72a76","Type":"ContainerStarted","Data":"edd9ed715cc82ede46dc11a8f6ce81dfe23ceea0f9a1d1b529905c9a49413456"} Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.232220 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" event={"ID":"3c0e7700-9f5b-4ece-8e9d-234c14d72a76","Type":"ContainerStarted","Data":"3e0e824189bbf24fd072a12cbfb36d79369cf99f60f91b94bfe893e592c3442c"} Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.232523 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.236328 4756 generic.go:334] "Generic (PLEG): container finished" podID="3021f6bf-d879-49e1-8051-9f599be91475" containerID="792e8e9b13b94648f06b09e7f420635b87c724a748da1cb9c2b117bfddf5683e" exitCode=0 Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.236432 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5999" event={"ID":"3021f6bf-d879-49e1-8051-9f599be91475","Type":"ContainerDied","Data":"792e8e9b13b94648f06b09e7f420635b87c724a748da1cb9c2b117bfddf5683e"} Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.236946 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5999" event={"ID":"3021f6bf-d879-49e1-8051-9f599be91475","Type":"ContainerStarted","Data":"9d826301fef4649b2b16b6c0ee1952067e20af8064b7cbb7912931ebf774b62c"} Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.239390 4756 generic.go:334] "Generic (PLEG): container finished" podID="4ad66a22-bd41-40e4-b88f-d222ff093a15" containerID="034d53bb0f792d650d2393313d8cea87998a2b9bc903232d0f83026b828531c2" exitCode=0 Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.239473 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvqhq" event={"ID":"4ad66a22-bd41-40e4-b88f-d222ff093a15","Type":"ContainerDied","Data":"034d53bb0f792d650d2393313d8cea87998a2b9bc903232d0f83026b828531c2"} Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.239509 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvqhq" event={"ID":"4ad66a22-bd41-40e4-b88f-d222ff093a15","Type":"ContainerStarted","Data":"f3b44280748c4f835e9f681e0f69b4e877f12e3cd1b1ea52b6881555ebec60fa"} Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.262464 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" podStartSLOduration=2.262424416 podStartE2EDuration="2.262424416s" podCreationTimestamp="2026-02-03 09:15:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:15:59.257658457 +0000 UTC m=+350.408125852" watchObservedRunningTime="2026-02-03 09:15:59.262424416 +0000 UTC m=+350.412891791" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.382618 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2nqkf"] Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.383627 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.385135 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.392563 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2nqkf"] Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.521882 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtqj9\" (UniqueName: \"kubernetes.io/projected/f507b140-0d1f-424b-8077-80c7ac9a1a15-kube-api-access-xtqj9\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.521939 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f507b140-0d1f-424b-8077-80c7ac9a1a15-utilities\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.522038 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f507b140-0d1f-424b-8077-80c7ac9a1a15-catalog-content\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.623254 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtqj9\" (UniqueName: \"kubernetes.io/projected/f507b140-0d1f-424b-8077-80c7ac9a1a15-kube-api-access-xtqj9\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.623297 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f507b140-0d1f-424b-8077-80c7ac9a1a15-utilities\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.623318 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f507b140-0d1f-424b-8077-80c7ac9a1a15-catalog-content\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.623777 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f507b140-0d1f-424b-8077-80c7ac9a1a15-catalog-content\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.623861 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f507b140-0d1f-424b-8077-80c7ac9a1a15-utilities\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.648767 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtqj9\" (UniqueName: \"kubernetes.io/projected/f507b140-0d1f-424b-8077-80c7ac9a1a15-kube-api-access-xtqj9\") pod \"community-operators-2nqkf\" (UID: \"f507b140-0d1f-424b-8077-80c7ac9a1a15\") " pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.697019 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:15:59 crc kubenswrapper[4756]: I0203 09:15:59.886614 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2nqkf"] Feb 03 09:15:59 crc kubenswrapper[4756]: W0203 09:15:59.891063 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf507b140_0d1f_424b_8077_80c7ac9a1a15.slice/crio-dcc3b8c45485c0751c2a411d4a867ca887cd31949d739b6f245e3173ea00a654 WatchSource:0}: Error finding container dcc3b8c45485c0751c2a411d4a867ca887cd31949d739b6f245e3173ea00a654: Status 404 returned error can't find the container with id dcc3b8c45485c0751c2a411d4a867ca887cd31949d739b6f245e3173ea00a654 Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.246327 4756 generic.go:334] "Generic (PLEG): container finished" podID="3021f6bf-d879-49e1-8051-9f599be91475" containerID="0452821a350366879607e1a03ad3250d5c71748bf022e15fdcc4116106d3af75" exitCode=0 Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.246384 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5999" event={"ID":"3021f6bf-d879-49e1-8051-9f599be91475","Type":"ContainerDied","Data":"0452821a350366879607e1a03ad3250d5c71748bf022e15fdcc4116106d3af75"} Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.249842 4756 generic.go:334] "Generic (PLEG): container finished" podID="f507b140-0d1f-424b-8077-80c7ac9a1a15" containerID="9728090d4cf8299e1a1cdfcf8522992371c80c1002ee263958e364146177bc43" exitCode=0 Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.250318 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2nqkf" event={"ID":"f507b140-0d1f-424b-8077-80c7ac9a1a15","Type":"ContainerDied","Data":"9728090d4cf8299e1a1cdfcf8522992371c80c1002ee263958e364146177bc43"} Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.250339 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2nqkf" event={"ID":"f507b140-0d1f-424b-8077-80c7ac9a1a15","Type":"ContainerStarted","Data":"dcc3b8c45485c0751c2a411d4a867ca887cd31949d739b6f245e3173ea00a654"} Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.786479 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kzdvb"] Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.787595 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.790520 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.799876 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kzdvb"] Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.838322 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-utilities\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.838384 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-catalog-content\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.838420 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd4sk\" (UniqueName: \"kubernetes.io/projected/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-kube-api-access-vd4sk\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.939605 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-utilities\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.939953 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-catalog-content\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.940068 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd4sk\" (UniqueName: \"kubernetes.io/projected/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-kube-api-access-vd4sk\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.940067 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-utilities\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.940386 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-catalog-content\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:00 crc kubenswrapper[4756]: I0203 09:16:00.962226 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd4sk\" (UniqueName: \"kubernetes.io/projected/74a5cfb6-5bda-4b5b-a828-1bf7a2705273-kube-api-access-vd4sk\") pod \"certified-operators-kzdvb\" (UID: \"74a5cfb6-5bda-4b5b-a828-1bf7a2705273\") " pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:01 crc kubenswrapper[4756]: I0203 09:16:01.107203 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:01 crc kubenswrapper[4756]: I0203 09:16:01.264915 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5999" event={"ID":"3021f6bf-d879-49e1-8051-9f599be91475","Type":"ContainerStarted","Data":"bf5105149244b6e27bb98684e663052323f216bb6db776cffb33e7549fe975ed"} Feb 03 09:16:01 crc kubenswrapper[4756]: I0203 09:16:01.266917 4756 generic.go:334] "Generic (PLEG): container finished" podID="4ad66a22-bd41-40e4-b88f-d222ff093a15" containerID="57d222eb1719dfbb1e01165b2a9062c2afe9bd9fa10253269044b48b3b045e6b" exitCode=0 Feb 03 09:16:01 crc kubenswrapper[4756]: I0203 09:16:01.266969 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvqhq" event={"ID":"4ad66a22-bd41-40e4-b88f-d222ff093a15","Type":"ContainerDied","Data":"57d222eb1719dfbb1e01165b2a9062c2afe9bd9fa10253269044b48b3b045e6b"} Feb 03 09:16:01 crc kubenswrapper[4756]: I0203 09:16:01.271818 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2nqkf" event={"ID":"f507b140-0d1f-424b-8077-80c7ac9a1a15","Type":"ContainerStarted","Data":"4f424e94ecc6b9650f1f51fd5ae79f36fd7a4537c81a82f85bbcdf17095d19fd"} Feb 03 09:16:01 crc kubenswrapper[4756]: I0203 09:16:01.291307 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h5999" podStartSLOduration=2.714476434 podStartE2EDuration="4.291291191s" podCreationTimestamp="2026-02-03 09:15:57 +0000 UTC" firstStartedPulling="2026-02-03 09:15:59.23830257 +0000 UTC m=+350.388769955" lastFinishedPulling="2026-02-03 09:16:00.815117337 +0000 UTC m=+351.965584712" observedRunningTime="2026-02-03 09:16:01.290405551 +0000 UTC m=+352.440872936" watchObservedRunningTime="2026-02-03 09:16:01.291291191 +0000 UTC m=+352.441758566" Feb 03 09:16:01 crc kubenswrapper[4756]: I0203 09:16:01.356570 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kzdvb"] Feb 03 09:16:01 crc kubenswrapper[4756]: W0203 09:16:01.374021 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74a5cfb6_5bda_4b5b_a828_1bf7a2705273.slice/crio-83c4fa1194b658b608d56812268aaadcdab7f515bbbe034d82c63e3a2383d466 WatchSource:0}: Error finding container 83c4fa1194b658b608d56812268aaadcdab7f515bbbe034d82c63e3a2383d466: Status 404 returned error can't find the container with id 83c4fa1194b658b608d56812268aaadcdab7f515bbbe034d82c63e3a2383d466 Feb 03 09:16:02 crc kubenswrapper[4756]: I0203 09:16:02.292914 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvqhq" event={"ID":"4ad66a22-bd41-40e4-b88f-d222ff093a15","Type":"ContainerStarted","Data":"b0cf1150b8f39632185b95e21cfe00fde68b793323ae96c68e0a081160578cd4"} Feb 03 09:16:02 crc kubenswrapper[4756]: I0203 09:16:02.295926 4756 generic.go:334] "Generic (PLEG): container finished" podID="74a5cfb6-5bda-4b5b-a828-1bf7a2705273" containerID="4d86f833d0e22afafcf9e7016018c82a94194e25d37beac157a4f8e32049deb4" exitCode=0 Feb 03 09:16:02 crc kubenswrapper[4756]: I0203 09:16:02.296000 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzdvb" event={"ID":"74a5cfb6-5bda-4b5b-a828-1bf7a2705273","Type":"ContainerDied","Data":"4d86f833d0e22afafcf9e7016018c82a94194e25d37beac157a4f8e32049deb4"} Feb 03 09:16:02 crc kubenswrapper[4756]: I0203 09:16:02.296029 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzdvb" event={"ID":"74a5cfb6-5bda-4b5b-a828-1bf7a2705273","Type":"ContainerStarted","Data":"83c4fa1194b658b608d56812268aaadcdab7f515bbbe034d82c63e3a2383d466"} Feb 03 09:16:02 crc kubenswrapper[4756]: I0203 09:16:02.302649 4756 generic.go:334] "Generic (PLEG): container finished" podID="f507b140-0d1f-424b-8077-80c7ac9a1a15" containerID="4f424e94ecc6b9650f1f51fd5ae79f36fd7a4537c81a82f85bbcdf17095d19fd" exitCode=0 Feb 03 09:16:02 crc kubenswrapper[4756]: I0203 09:16:02.303996 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2nqkf" event={"ID":"f507b140-0d1f-424b-8077-80c7ac9a1a15","Type":"ContainerDied","Data":"4f424e94ecc6b9650f1f51fd5ae79f36fd7a4537c81a82f85bbcdf17095d19fd"} Feb 03 09:16:02 crc kubenswrapper[4756]: I0203 09:16:02.342330 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dvqhq" podStartSLOduration=2.920731696 podStartE2EDuration="5.342292194s" podCreationTimestamp="2026-02-03 09:15:57 +0000 UTC" firstStartedPulling="2026-02-03 09:15:59.2415881 +0000 UTC m=+350.392055485" lastFinishedPulling="2026-02-03 09:16:01.663148608 +0000 UTC m=+352.813615983" observedRunningTime="2026-02-03 09:16:02.320168755 +0000 UTC m=+353.470636130" watchObservedRunningTime="2026-02-03 09:16:02.342292194 +0000 UTC m=+353.492759579" Feb 03 09:16:03 crc kubenswrapper[4756]: I0203 09:16:03.318541 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2nqkf" event={"ID":"f507b140-0d1f-424b-8077-80c7ac9a1a15","Type":"ContainerStarted","Data":"07bbdaf797866e7394a9a73e200016876f21bbe2f104450ea5e8f4a2497b6191"} Feb 03 09:16:03 crc kubenswrapper[4756]: I0203 09:16:03.323213 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzdvb" event={"ID":"74a5cfb6-5bda-4b5b-a828-1bf7a2705273","Type":"ContainerStarted","Data":"06887641588a5b7815b380709ecebc811cf06d4f9238258cc55d1471273bb503"} Feb 03 09:16:03 crc kubenswrapper[4756]: I0203 09:16:03.345300 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2nqkf" podStartSLOduration=1.754625094 podStartE2EDuration="4.345283864s" podCreationTimestamp="2026-02-03 09:15:59 +0000 UTC" firstStartedPulling="2026-02-03 09:16:00.251248732 +0000 UTC m=+351.401716107" lastFinishedPulling="2026-02-03 09:16:02.841907502 +0000 UTC m=+353.992374877" observedRunningTime="2026-02-03 09:16:03.344063583 +0000 UTC m=+354.494530978" watchObservedRunningTime="2026-02-03 09:16:03.345283864 +0000 UTC m=+354.495751239" Feb 03 09:16:04 crc kubenswrapper[4756]: I0203 09:16:04.329055 4756 generic.go:334] "Generic (PLEG): container finished" podID="74a5cfb6-5bda-4b5b-a828-1bf7a2705273" containerID="06887641588a5b7815b380709ecebc811cf06d4f9238258cc55d1471273bb503" exitCode=0 Feb 03 09:16:04 crc kubenswrapper[4756]: I0203 09:16:04.329128 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzdvb" event={"ID":"74a5cfb6-5bda-4b5b-a828-1bf7a2705273","Type":"ContainerDied","Data":"06887641588a5b7815b380709ecebc811cf06d4f9238258cc55d1471273bb503"} Feb 03 09:16:05 crc kubenswrapper[4756]: I0203 09:16:05.336612 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzdvb" event={"ID":"74a5cfb6-5bda-4b5b-a828-1bf7a2705273","Type":"ContainerStarted","Data":"97767ec9060f24875cc330dd8baca916fcfeb5fd7a85cc7fbf93c7836091a03e"} Feb 03 09:16:05 crc kubenswrapper[4756]: I0203 09:16:05.362493 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kzdvb" podStartSLOduration=2.936042647 podStartE2EDuration="5.362476188s" podCreationTimestamp="2026-02-03 09:16:00 +0000 UTC" firstStartedPulling="2026-02-03 09:16:02.298677027 +0000 UTC m=+353.449144402" lastFinishedPulling="2026-02-03 09:16:04.725110568 +0000 UTC m=+355.875577943" observedRunningTime="2026-02-03 09:16:05.361274698 +0000 UTC m=+356.511742083" watchObservedRunningTime="2026-02-03 09:16:05.362476188 +0000 UTC m=+356.512943573" Feb 03 09:16:07 crc kubenswrapper[4756]: I0203 09:16:07.906559 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:16:07 crc kubenswrapper[4756]: I0203 09:16:07.906923 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:16:07 crc kubenswrapper[4756]: I0203 09:16:07.977635 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:16:08 crc kubenswrapper[4756]: I0203 09:16:08.100797 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:16:08 crc kubenswrapper[4756]: I0203 09:16:08.100876 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:16:08 crc kubenswrapper[4756]: I0203 09:16:08.169530 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:16:08 crc kubenswrapper[4756]: I0203 09:16:08.404254 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h5999" Feb 03 09:16:08 crc kubenswrapper[4756]: I0203 09:16:08.406155 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dvqhq" Feb 03 09:16:09 crc kubenswrapper[4756]: I0203 09:16:09.697683 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:16:09 crc kubenswrapper[4756]: I0203 09:16:09.698422 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:16:09 crc kubenswrapper[4756]: I0203 09:16:09.747658 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:16:10 crc kubenswrapper[4756]: I0203 09:16:10.428915 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2nqkf" Feb 03 09:16:11 crc kubenswrapper[4756]: I0203 09:16:11.108195 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:11 crc kubenswrapper[4756]: I0203 09:16:11.109257 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:11 crc kubenswrapper[4756]: I0203 09:16:11.155350 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:11 crc kubenswrapper[4756]: I0203 09:16:11.404366 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kzdvb" Feb 03 09:16:13 crc kubenswrapper[4756]: I0203 09:16:13.566400 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:16:13 crc kubenswrapper[4756]: I0203 09:16:13.567603 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:16:18 crc kubenswrapper[4756]: I0203 09:16:18.309510 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-r8qjq" Feb 03 09:16:18 crc kubenswrapper[4756]: I0203 09:16:18.375973 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxxsv"] Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.418622 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" podUID="e579b706-5cc4-4a92-94f2-b22ad2135ffc" containerName="registry" containerID="cri-o://3fa1da28099b8974f60dbf71def0ff3fd5c772f1ef99b6af2a6ad2dfd126e834" gracePeriod=30 Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.546611 4756 generic.go:334] "Generic (PLEG): container finished" podID="e579b706-5cc4-4a92-94f2-b22ad2135ffc" containerID="3fa1da28099b8974f60dbf71def0ff3fd5c772f1ef99b6af2a6ad2dfd126e834" exitCode=0 Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.546685 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" event={"ID":"e579b706-5cc4-4a92-94f2-b22ad2135ffc","Type":"ContainerDied","Data":"3fa1da28099b8974f60dbf71def0ff3fd5c772f1ef99b6af2a6ad2dfd126e834"} Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.566675 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.567019 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.567073 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.567746 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a749937f8614549763f69387e47ae3407826bd178fc787c90a020ffe22fb54e1"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.567818 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://a749937f8614549763f69387e47ae3407826bd178fc787c90a020ffe22fb54e1" gracePeriod=600 Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.793087 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.837246 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.837735 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-tls\") pod \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.837758 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-trusted-ca\") pod \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.837792 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-bound-sa-token\") pod \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.837824 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtdsg\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-kube-api-access-dtdsg\") pod \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.837863 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-certificates\") pod \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.838968 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e579b706-5cc4-4a92-94f2-b22ad2135ffc-ca-trust-extracted\") pod \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.838981 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e579b706-5cc4-4a92-94f2-b22ad2135ffc" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.838999 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e579b706-5cc4-4a92-94f2-b22ad2135ffc" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.839028 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e579b706-5cc4-4a92-94f2-b22ad2135ffc-installation-pull-secrets\") pod \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\" (UID: \"e579b706-5cc4-4a92-94f2-b22ad2135ffc\") " Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.839256 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.839274 4756 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.847078 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e579b706-5cc4-4a92-94f2-b22ad2135ffc" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.847296 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-kube-api-access-dtdsg" (OuterVolumeSpecName: "kube-api-access-dtdsg") pod "e579b706-5cc4-4a92-94f2-b22ad2135ffc" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc"). InnerVolumeSpecName "kube-api-access-dtdsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.847565 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e579b706-5cc4-4a92-94f2-b22ad2135ffc" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.847750 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e579b706-5cc4-4a92-94f2-b22ad2135ffc-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e579b706-5cc4-4a92-94f2-b22ad2135ffc" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.849835 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e579b706-5cc4-4a92-94f2-b22ad2135ffc" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.860905 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e579b706-5cc4-4a92-94f2-b22ad2135ffc-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e579b706-5cc4-4a92-94f2-b22ad2135ffc" (UID: "e579b706-5cc4-4a92-94f2-b22ad2135ffc"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.940605 4756 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.940655 4756 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.940667 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtdsg\" (UniqueName: \"kubernetes.io/projected/e579b706-5cc4-4a92-94f2-b22ad2135ffc-kube-api-access-dtdsg\") on node \"crc\" DevicePath \"\"" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.940680 4756 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e579b706-5cc4-4a92-94f2-b22ad2135ffc-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 03 09:16:43 crc kubenswrapper[4756]: I0203 09:16:43.940691 4756 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e579b706-5cc4-4a92-94f2-b22ad2135ffc-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.553305 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="a749937f8614549763f69387e47ae3407826bd178fc787c90a020ffe22fb54e1" exitCode=0 Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.553376 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"a749937f8614549763f69387e47ae3407826bd178fc787c90a020ffe22fb54e1"} Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.553410 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"47fef4302ae2be888f70b5160d00aea824e230accaf11fd9a2599e23eb41dc66"} Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.553435 4756 scope.go:117] "RemoveContainer" containerID="df0307859813e9f8c558d091977a2c97d1b0679d9071772af1a8a1d467ca6872" Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.555824 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" event={"ID":"e579b706-5cc4-4a92-94f2-b22ad2135ffc","Type":"ContainerDied","Data":"fe46f7277a65b7646abd62fc32a5c9d6ea9a3c924f57dbd7346feef9304a4771"} Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.555882 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxxsv" Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.581427 4756 scope.go:117] "RemoveContainer" containerID="3fa1da28099b8974f60dbf71def0ff3fd5c772f1ef99b6af2a6ad2dfd126e834" Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.594179 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxxsv"] Feb 03 09:16:44 crc kubenswrapper[4756]: I0203 09:16:44.596538 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxxsv"] Feb 03 09:16:45 crc kubenswrapper[4756]: I0203 09:16:45.624187 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e579b706-5cc4-4a92-94f2-b22ad2135ffc" path="/var/lib/kubelet/pods/e579b706-5cc4-4a92-94f2-b22ad2135ffc/volumes" Feb 03 09:18:43 crc kubenswrapper[4756]: I0203 09:18:43.566257 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:18:43 crc kubenswrapper[4756]: I0203 09:18:43.567011 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:19:09 crc kubenswrapper[4756]: I0203 09:19:09.844224 4756 scope.go:117] "RemoveContainer" containerID="86c4adbb56c70b5e3b7cbab7f5522d2218131fe50218672931f616f916548e18" Feb 03 09:19:09 crc kubenswrapper[4756]: I0203 09:19:09.877405 4756 scope.go:117] "RemoveContainer" containerID="5a013ce0da93436a9256ffa45255470f734a9b1e895a504aec1decdd199d08a5" Feb 03 09:19:13 crc kubenswrapper[4756]: I0203 09:19:13.565666 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:19:13 crc kubenswrapper[4756]: I0203 09:19:13.565971 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:19:43 crc kubenswrapper[4756]: I0203 09:19:43.566619 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:19:43 crc kubenswrapper[4756]: I0203 09:19:43.568718 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:19:43 crc kubenswrapper[4756]: I0203 09:19:43.568826 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:19:43 crc kubenswrapper[4756]: I0203 09:19:43.570266 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"47fef4302ae2be888f70b5160d00aea824e230accaf11fd9a2599e23eb41dc66"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:19:43 crc kubenswrapper[4756]: I0203 09:19:43.570406 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://47fef4302ae2be888f70b5160d00aea824e230accaf11fd9a2599e23eb41dc66" gracePeriod=600 Feb 03 09:19:44 crc kubenswrapper[4756]: I0203 09:19:44.568720 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="47fef4302ae2be888f70b5160d00aea824e230accaf11fd9a2599e23eb41dc66" exitCode=0 Feb 03 09:19:44 crc kubenswrapper[4756]: I0203 09:19:44.568805 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"47fef4302ae2be888f70b5160d00aea824e230accaf11fd9a2599e23eb41dc66"} Feb 03 09:19:44 crc kubenswrapper[4756]: I0203 09:19:44.569036 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"c20ecab788892a342dab2838ffafa76a4cbfbef0fe19794f920df87b66dcf0b1"} Feb 03 09:19:44 crc kubenswrapper[4756]: I0203 09:19:44.569056 4756 scope.go:117] "RemoveContainer" containerID="a749937f8614549763f69387e47ae3407826bd178fc787c90a020ffe22fb54e1" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.196071 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj"] Feb 03 09:21:41 crc kubenswrapper[4756]: E0203 09:21:41.196899 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e579b706-5cc4-4a92-94f2-b22ad2135ffc" containerName="registry" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.196916 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e579b706-5cc4-4a92-94f2-b22ad2135ffc" containerName="registry" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.197042 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="e579b706-5cc4-4a92-94f2-b22ad2135ffc" containerName="registry" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.197492 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.199842 4756 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-s5wq5" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.200028 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.200184 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.200364 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-fm5vv"] Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.201122 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-fm5vv" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.203064 4756 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-76jfc" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.214249 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj"] Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.255027 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsq28\" (UniqueName: \"kubernetes.io/projected/e8c43b40-0d26-4582-8ffd-b2387527edd6-kube-api-access-xsq28\") pod \"cert-manager-858654f9db-fm5vv\" (UID: \"e8c43b40-0d26-4582-8ffd-b2387527edd6\") " pod="cert-manager/cert-manager-858654f9db-fm5vv" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.255109 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl7m9\" (UniqueName: \"kubernetes.io/projected/2c537fcd-648d-4fea-ab9a-0e0edfa8b14b-kube-api-access-pl7m9\") pod \"cert-manager-cainjector-cf98fcc89-gnvfj\" (UID: \"2c537fcd-648d-4fea-ab9a-0e0edfa8b14b\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.270036 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-fm5vv"] Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.270130 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-gnr5k"] Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.271009 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.273916 4756 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-5p29d" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.274924 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-gnr5k"] Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.356458 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkd4p\" (UniqueName: \"kubernetes.io/projected/a4b8e083-79fc-4bbe-b3c8-16e237b63458-kube-api-access-zkd4p\") pod \"cert-manager-webhook-687f57d79b-gnr5k\" (UID: \"a4b8e083-79fc-4bbe-b3c8-16e237b63458\") " pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.356554 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsq28\" (UniqueName: \"kubernetes.io/projected/e8c43b40-0d26-4582-8ffd-b2387527edd6-kube-api-access-xsq28\") pod \"cert-manager-858654f9db-fm5vv\" (UID: \"e8c43b40-0d26-4582-8ffd-b2387527edd6\") " pod="cert-manager/cert-manager-858654f9db-fm5vv" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.356589 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl7m9\" (UniqueName: \"kubernetes.io/projected/2c537fcd-648d-4fea-ab9a-0e0edfa8b14b-kube-api-access-pl7m9\") pod \"cert-manager-cainjector-cf98fcc89-gnvfj\" (UID: \"2c537fcd-648d-4fea-ab9a-0e0edfa8b14b\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.376365 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl7m9\" (UniqueName: \"kubernetes.io/projected/2c537fcd-648d-4fea-ab9a-0e0edfa8b14b-kube-api-access-pl7m9\") pod \"cert-manager-cainjector-cf98fcc89-gnvfj\" (UID: \"2c537fcd-648d-4fea-ab9a-0e0edfa8b14b\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.378679 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsq28\" (UniqueName: \"kubernetes.io/projected/e8c43b40-0d26-4582-8ffd-b2387527edd6-kube-api-access-xsq28\") pod \"cert-manager-858654f9db-fm5vv\" (UID: \"e8c43b40-0d26-4582-8ffd-b2387527edd6\") " pod="cert-manager/cert-manager-858654f9db-fm5vv" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.457427 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkd4p\" (UniqueName: \"kubernetes.io/projected/a4b8e083-79fc-4bbe-b3c8-16e237b63458-kube-api-access-zkd4p\") pod \"cert-manager-webhook-687f57d79b-gnr5k\" (UID: \"a4b8e083-79fc-4bbe-b3c8-16e237b63458\") " pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.498103 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkd4p\" (UniqueName: \"kubernetes.io/projected/a4b8e083-79fc-4bbe-b3c8-16e237b63458-kube-api-access-zkd4p\") pod \"cert-manager-webhook-687f57d79b-gnr5k\" (UID: \"a4b8e083-79fc-4bbe-b3c8-16e237b63458\") " pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.570175 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.570953 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-fm5vv" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.591338 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.821556 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj"] Feb 03 09:21:41 crc kubenswrapper[4756]: I0203 09:21:41.832687 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 09:21:42 crc kubenswrapper[4756]: I0203 09:21:42.094840 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-gnr5k"] Feb 03 09:21:42 crc kubenswrapper[4756]: W0203 09:21:42.096034 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4b8e083_79fc_4bbe_b3c8_16e237b63458.slice/crio-8106f64b5cead3742100a32889cf51b5509b71aa987020cd5d24fcfffc94d38d WatchSource:0}: Error finding container 8106f64b5cead3742100a32889cf51b5509b71aa987020cd5d24fcfffc94d38d: Status 404 returned error can't find the container with id 8106f64b5cead3742100a32889cf51b5509b71aa987020cd5d24fcfffc94d38d Feb 03 09:21:42 crc kubenswrapper[4756]: I0203 09:21:42.103516 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-fm5vv"] Feb 03 09:21:42 crc kubenswrapper[4756]: I0203 09:21:42.212407 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj" event={"ID":"2c537fcd-648d-4fea-ab9a-0e0edfa8b14b","Type":"ContainerStarted","Data":"d17b3b9972e35e2aaaeecaa8f6ed2f52021b5989d047ace06020413909b2e6a8"} Feb 03 09:21:42 crc kubenswrapper[4756]: I0203 09:21:42.213626 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-fm5vv" event={"ID":"e8c43b40-0d26-4582-8ffd-b2387527edd6","Type":"ContainerStarted","Data":"04f13c4e2f652877b36ccf60b7fa18f47985c2ab0b5e27d13081b309bf5f31d0"} Feb 03 09:21:42 crc kubenswrapper[4756]: I0203 09:21:42.214662 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" event={"ID":"a4b8e083-79fc-4bbe-b3c8-16e237b63458","Type":"ContainerStarted","Data":"8106f64b5cead3742100a32889cf51b5509b71aa987020cd5d24fcfffc94d38d"} Feb 03 09:21:43 crc kubenswrapper[4756]: I0203 09:21:43.566489 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:21:43 crc kubenswrapper[4756]: I0203 09:21:43.566543 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:21:44 crc kubenswrapper[4756]: I0203 09:21:44.227857 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj" event={"ID":"2c537fcd-648d-4fea-ab9a-0e0edfa8b14b","Type":"ContainerStarted","Data":"70d560e50f9c3d5b10797353a2879236162be0810b055f90f7d180a829991d59"} Feb 03 09:21:46 crc kubenswrapper[4756]: I0203 09:21:46.252987 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-fm5vv" event={"ID":"e8c43b40-0d26-4582-8ffd-b2387527edd6","Type":"ContainerStarted","Data":"aa19d24d645c65d95aef884f79de13fcb865f6da3ad64d2653845e780a1e7975"} Feb 03 09:21:46 crc kubenswrapper[4756]: I0203 09:21:46.255754 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" event={"ID":"a4b8e083-79fc-4bbe-b3c8-16e237b63458","Type":"ContainerStarted","Data":"768c53ddcb110cb819af05aa345ea1e5a9e7ab3366dd2ad2494e46784cceaba1"} Feb 03 09:21:46 crc kubenswrapper[4756]: I0203 09:21:46.255919 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" Feb 03 09:21:46 crc kubenswrapper[4756]: I0203 09:21:46.275879 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gnvfj" podStartSLOduration=3.306577851 podStartE2EDuration="5.275859572s" podCreationTimestamp="2026-02-03 09:21:41 +0000 UTC" firstStartedPulling="2026-02-03 09:21:41.832471651 +0000 UTC m=+692.982939026" lastFinishedPulling="2026-02-03 09:21:43.801753372 +0000 UTC m=+694.952220747" observedRunningTime="2026-02-03 09:21:44.249048265 +0000 UTC m=+695.399515660" watchObservedRunningTime="2026-02-03 09:21:46.275859572 +0000 UTC m=+697.426326947" Feb 03 09:21:46 crc kubenswrapper[4756]: I0203 09:21:46.280946 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-fm5vv" podStartSLOduration=2.268394823 podStartE2EDuration="5.28092104s" podCreationTimestamp="2026-02-03 09:21:41 +0000 UTC" firstStartedPulling="2026-02-03 09:21:42.105157712 +0000 UTC m=+693.255625097" lastFinishedPulling="2026-02-03 09:21:45.117683929 +0000 UTC m=+696.268151314" observedRunningTime="2026-02-03 09:21:46.272576699 +0000 UTC m=+697.423044074" watchObservedRunningTime="2026-02-03 09:21:46.28092104 +0000 UTC m=+697.431388435" Feb 03 09:21:46 crc kubenswrapper[4756]: I0203 09:21:46.293681 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" podStartSLOduration=2.224541894 podStartE2EDuration="5.293662537s" podCreationTimestamp="2026-02-03 09:21:41 +0000 UTC" firstStartedPulling="2026-02-03 09:21:42.098241527 +0000 UTC m=+693.248708902" lastFinishedPulling="2026-02-03 09:21:45.16736217 +0000 UTC m=+696.317829545" observedRunningTime="2026-02-03 09:21:46.292242183 +0000 UTC m=+697.442709658" watchObservedRunningTime="2026-02-03 09:21:46.293662537 +0000 UTC m=+697.444129922" Feb 03 09:21:51 crc kubenswrapper[4756]: I0203 09:21:51.594546 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-gnr5k" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.506917 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gfsqr"] Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.507825 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovn-controller" containerID="cri-o://7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391" gracePeriod=30 Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.508237 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="sbdb" containerID="cri-o://c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" gracePeriod=30 Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.508290 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="nbdb" containerID="cri-o://1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" gracePeriod=30 Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.508320 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="northd" containerID="cri-o://2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a" gracePeriod=30 Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.508347 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b" gracePeriod=30 Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.508375 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kube-rbac-proxy-node" containerID="cri-o://e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90" gracePeriod=30 Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.508402 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovn-acl-logging" containerID="cri-o://8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb" gracePeriod=30 Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.540372 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" containerID="cri-o://8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" gracePeriod=30 Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.683519 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.685197 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.686286 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.686323 4756 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="sbdb" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.687146 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.688438 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.690865 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.690913 4756 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="nbdb" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.858593 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/3.log" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.861125 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovn-acl-logging/0.log" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.861525 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovn-controller/0.log" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.861968 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914113 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vlrv6"] Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914300 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovn-acl-logging" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914311 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovn-acl-logging" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914319 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914325 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914335 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kubecfg-setup" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914341 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kubecfg-setup" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914350 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovn-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914356 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovn-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914367 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kube-rbac-proxy-node" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914373 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kube-rbac-proxy-node" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914382 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914387 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914394 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914400 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914408 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kube-rbac-proxy-ovn-metrics" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914415 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kube-rbac-proxy-ovn-metrics" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914422 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914427 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914453 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="sbdb" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914460 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="sbdb" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914467 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914473 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914481 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="nbdb" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914488 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="nbdb" Feb 03 09:21:59 crc kubenswrapper[4756]: E0203 09:21:59.914496 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="northd" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914502 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="northd" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914589 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914598 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovn-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914605 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="sbdb" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914611 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kube-rbac-proxy-node" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914623 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914630 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovn-acl-logging" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914637 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914644 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914653 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="northd" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914659 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="kube-rbac-proxy-ovn-metrics" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914666 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="nbdb" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.914821 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerName="ovnkube-controller" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.916257 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.980815 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-config\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.980884 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h849g\" (UniqueName: \"kubernetes.io/projected/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-kube-api-access-h849g\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.980934 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-script-lib\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.980962 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-systemd\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981003 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-slash\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981036 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-slash" (OuterVolumeSpecName: "host-slash") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981097 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-netns\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981134 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981541 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981684 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-ovn\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981724 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981741 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-bin\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981759 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-netd\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981789 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-var-lib-openvswitch\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981814 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981829 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-systemd-units\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981834 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981846 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-log-socket\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981852 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981864 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-node-log\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981865 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981860 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981879 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981891 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-ovn-kubernetes\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981900 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-node-log" (OuterVolumeSpecName: "node-log") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981903 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-log-socket" (OuterVolumeSpecName: "log-socket") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981914 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovn-node-metrics-cert\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981954 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-kubelet\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981969 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.981988 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-env-overrides\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982000 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982017 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-openvswitch\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982039 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-etc-openvswitch\") pod \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\" (UID: \"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705\") " Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982047 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982118 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fc550307-2c84-4d4f-ae1a-e415be21241d-ovn-node-metrics-cert\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982129 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982149 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-ovn\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982197 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-env-overrides\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982221 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-node-log\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982254 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-ovnkube-script-lib\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982283 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-run-ovn-kubernetes\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982306 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-etc-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982310 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982329 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-ovnkube-config\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982391 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-var-lib-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982471 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-slash\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982501 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-systemd\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982531 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-systemd-units\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982557 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982590 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982629 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-cni-bin\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982650 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-kubelet\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982672 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-cni-netd\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982720 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-run-netns\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982762 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrwnt\" (UniqueName: \"kubernetes.io/projected/fc550307-2c84-4d4f-ae1a-e415be21241d-kube-api-access-xrwnt\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982794 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-log-socket\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982922 4756 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982936 4756 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982948 4756 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982961 4756 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-slash\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982971 4756 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982983 4756 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.982994 4756 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983003 4756 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983013 4756 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983024 4756 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983035 4756 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983045 4756 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-log-socket\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983055 4756 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-node-log\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983066 4756 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983075 4756 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983084 4756 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.983467 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.986686 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-kube-api-access-h849g" (OuterVolumeSpecName: "kube-api-access-h849g") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "kube-api-access-h849g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.987060 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:21:59 crc kubenswrapper[4756]: I0203 09:21:59.994801 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" (UID: "3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.083981 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-run-netns\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084039 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrwnt\" (UniqueName: \"kubernetes.io/projected/fc550307-2c84-4d4f-ae1a-e415be21241d-kube-api-access-xrwnt\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084059 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-log-socket\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084086 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fc550307-2c84-4d4f-ae1a-e415be21241d-ovn-node-metrics-cert\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084107 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-ovn\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084138 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-env-overrides\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084132 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-run-netns\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084155 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-log-socket\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084155 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-node-log\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084191 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-node-log\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084211 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-ovn\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084224 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-ovnkube-script-lib\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084246 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-run-ovn-kubernetes\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084270 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-etc-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084291 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-ovnkube-config\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084308 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-var-lib-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084322 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-slash\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084337 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-systemd\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084355 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-systemd-units\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084378 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084412 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084438 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-var-lib-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084521 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-systemd-units\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084548 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-cni-bin\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084588 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084594 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-run-systemd\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084619 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084557 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-slash\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084657 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-kubelet\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084625 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-cni-bin\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084670 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-run-ovn-kubernetes\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084687 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-kubelet\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084698 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-etc-openvswitch\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084750 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-cni-netd\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084785 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fc550307-2c84-4d4f-ae1a-e415be21241d-host-cni-netd\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084826 4756 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084841 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h849g\" (UniqueName: \"kubernetes.io/projected/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-kube-api-access-h849g\") on node \"crc\" DevicePath \"\"" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084850 4756 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.084859 4756 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.085096 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-env-overrides\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.085191 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-ovnkube-config\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.085226 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fc550307-2c84-4d4f-ae1a-e415be21241d-ovnkube-script-lib\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.087161 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fc550307-2c84-4d4f-ae1a-e415be21241d-ovn-node-metrics-cert\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.098673 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrwnt\" (UniqueName: \"kubernetes.io/projected/fc550307-2c84-4d4f-ae1a-e415be21241d-kube-api-access-xrwnt\") pod \"ovnkube-node-vlrv6\" (UID: \"fc550307-2c84-4d4f-ae1a-e415be21241d\") " pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.232156 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.337179 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovnkube-controller/3.log" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.340184 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovn-acl-logging/0.log" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.340682 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gfsqr_3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/ovn-controller/0.log" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341062 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" exitCode=0 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341092 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" exitCode=0 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341101 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" exitCode=0 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341111 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a" exitCode=0 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341119 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b" exitCode=0 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341128 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90" exitCode=0 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341136 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb" exitCode=143 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341145 4756 generic.go:334] "Generic (PLEG): container finished" podID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" containerID="7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391" exitCode=143 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341152 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341141 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341222 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341238 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341245 4756 scope.go:117] "RemoveContainer" containerID="8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341250 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341346 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341372 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341387 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341400 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341408 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341414 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341420 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341424 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341430 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341435 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341459 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341476 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341495 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341509 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341515 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341521 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341526 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341531 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341537 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341542 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341548 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341553 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341561 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341570 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341576 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341581 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341586 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341591 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341597 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341602 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341607 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341612 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341619 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341626 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfsqr" event={"ID":"3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705","Type":"ContainerDied","Data":"28fa062ddd59b44435884484b416fcd11229a372d50d1998c980b048fdb703eb"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341634 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341639 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341644 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341649 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341654 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341659 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341664 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341669 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341674 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.341678 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.343712 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/2.log" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.344032 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/1.log" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.344060 4756 generic.go:334] "Generic (PLEG): container finished" podID="89e5d60f-f690-486f-b078-5ae9f98e1f3a" containerID="b0a499eb23a8e69d38e737205e6ed33a99d4a789663122520887ad03b0c3fb57" exitCode=2 Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.344132 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6srkq" event={"ID":"89e5d60f-f690-486f-b078-5ae9f98e1f3a","Type":"ContainerDied","Data":"b0a499eb23a8e69d38e737205e6ed33a99d4a789663122520887ad03b0c3fb57"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.344147 4756 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.344599 4756 scope.go:117] "RemoveContainer" containerID="b0a499eb23a8e69d38e737205e6ed33a99d4a789663122520887ad03b0c3fb57" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.344858 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6srkq_openshift-multus(89e5d60f-f690-486f-b078-5ae9f98e1f3a)\"" pod="openshift-multus/multus-6srkq" podUID="89e5d60f-f690-486f-b078-5ae9f98e1f3a" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.346379 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"78f31a8b9927dab52fcce8a0b5aa5abfee2d6e705146b85b02190d18a31936de"} Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.360642 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.386382 4756 scope.go:117] "RemoveContainer" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.399071 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gfsqr"] Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.402016 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gfsqr"] Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.440006 4756 scope.go:117] "RemoveContainer" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.453946 4756 scope.go:117] "RemoveContainer" containerID="2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.465470 4756 scope.go:117] "RemoveContainer" containerID="7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.487622 4756 scope.go:117] "RemoveContainer" containerID="e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.504270 4756 scope.go:117] "RemoveContainer" containerID="8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.519067 4756 scope.go:117] "RemoveContainer" containerID="7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.531074 4756 scope.go:117] "RemoveContainer" containerID="7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.543302 4756 scope.go:117] "RemoveContainer" containerID="8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.543740 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": container with ID starting with 8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4 not found: ID does not exist" containerID="8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.543794 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} err="failed to get container status \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": rpc error: code = NotFound desc = could not find container \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": container with ID starting with 8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.543818 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.544106 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": container with ID starting with f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b not found: ID does not exist" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.544147 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} err="failed to get container status \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": rpc error: code = NotFound desc = could not find container \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": container with ID starting with f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.544174 4756 scope.go:117] "RemoveContainer" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.544493 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": container with ID starting with c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f not found: ID does not exist" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.544541 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} err="failed to get container status \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": rpc error: code = NotFound desc = could not find container \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": container with ID starting with c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.544568 4756 scope.go:117] "RemoveContainer" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.544840 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": container with ID starting with 1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815 not found: ID does not exist" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.544873 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} err="failed to get container status \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": rpc error: code = NotFound desc = could not find container \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": container with ID starting with 1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.544896 4756 scope.go:117] "RemoveContainer" containerID="2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.545112 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": container with ID starting with 2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a not found: ID does not exist" containerID="2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.545135 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} err="failed to get container status \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": rpc error: code = NotFound desc = could not find container \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": container with ID starting with 2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.545148 4756 scope.go:117] "RemoveContainer" containerID="7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.545387 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": container with ID starting with 7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b not found: ID does not exist" containerID="7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.545412 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} err="failed to get container status \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": rpc error: code = NotFound desc = could not find container \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": container with ID starting with 7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.545433 4756 scope.go:117] "RemoveContainer" containerID="e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.545701 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": container with ID starting with e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90 not found: ID does not exist" containerID="e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.545724 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} err="failed to get container status \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": rpc error: code = NotFound desc = could not find container \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": container with ID starting with e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.545737 4756 scope.go:117] "RemoveContainer" containerID="8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.545998 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": container with ID starting with 8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb not found: ID does not exist" containerID="8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.546033 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} err="failed to get container status \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": rpc error: code = NotFound desc = could not find container \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": container with ID starting with 8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.546051 4756 scope.go:117] "RemoveContainer" containerID="7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.546259 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": container with ID starting with 7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391 not found: ID does not exist" containerID="7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.546279 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} err="failed to get container status \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": rpc error: code = NotFound desc = could not find container \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": container with ID starting with 7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.546298 4756 scope.go:117] "RemoveContainer" containerID="7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc" Feb 03 09:22:00 crc kubenswrapper[4756]: E0203 09:22:00.546696 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": container with ID starting with 7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc not found: ID does not exist" containerID="7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.546728 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} err="failed to get container status \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": rpc error: code = NotFound desc = could not find container \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": container with ID starting with 7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.546745 4756 scope.go:117] "RemoveContainer" containerID="8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.546933 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} err="failed to get container status \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": rpc error: code = NotFound desc = could not find container \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": container with ID starting with 8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.546955 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.547171 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} err="failed to get container status \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": rpc error: code = NotFound desc = could not find container \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": container with ID starting with f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.547213 4756 scope.go:117] "RemoveContainer" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.547402 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} err="failed to get container status \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": rpc error: code = NotFound desc = could not find container \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": container with ID starting with c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.547426 4756 scope.go:117] "RemoveContainer" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.547650 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} err="failed to get container status \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": rpc error: code = NotFound desc = could not find container \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": container with ID starting with 1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.547669 4756 scope.go:117] "RemoveContainer" containerID="2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.548088 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} err="failed to get container status \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": rpc error: code = NotFound desc = could not find container \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": container with ID starting with 2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.548138 4756 scope.go:117] "RemoveContainer" containerID="7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.548397 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} err="failed to get container status \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": rpc error: code = NotFound desc = could not find container \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": container with ID starting with 7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.548420 4756 scope.go:117] "RemoveContainer" containerID="e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.548639 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} err="failed to get container status \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": rpc error: code = NotFound desc = could not find container \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": container with ID starting with e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.548662 4756 scope.go:117] "RemoveContainer" containerID="8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.548838 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} err="failed to get container status \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": rpc error: code = NotFound desc = could not find container \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": container with ID starting with 8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.548862 4756 scope.go:117] "RemoveContainer" containerID="7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549098 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} err="failed to get container status \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": rpc error: code = NotFound desc = could not find container \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": container with ID starting with 7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549121 4756 scope.go:117] "RemoveContainer" containerID="7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549282 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} err="failed to get container status \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": rpc error: code = NotFound desc = could not find container \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": container with ID starting with 7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549301 4756 scope.go:117] "RemoveContainer" containerID="8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549568 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} err="failed to get container status \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": rpc error: code = NotFound desc = could not find container \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": container with ID starting with 8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549588 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549773 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} err="failed to get container status \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": rpc error: code = NotFound desc = could not find container \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": container with ID starting with f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549792 4756 scope.go:117] "RemoveContainer" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549963 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} err="failed to get container status \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": rpc error: code = NotFound desc = could not find container \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": container with ID starting with c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.549990 4756 scope.go:117] "RemoveContainer" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.550237 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} err="failed to get container status \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": rpc error: code = NotFound desc = could not find container \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": container with ID starting with 1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.550256 4756 scope.go:117] "RemoveContainer" containerID="2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.550471 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} err="failed to get container status \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": rpc error: code = NotFound desc = could not find container \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": container with ID starting with 2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.550491 4756 scope.go:117] "RemoveContainer" containerID="7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.550684 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} err="failed to get container status \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": rpc error: code = NotFound desc = could not find container \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": container with ID starting with 7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.550703 4756 scope.go:117] "RemoveContainer" containerID="e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.550872 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} err="failed to get container status \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": rpc error: code = NotFound desc = could not find container \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": container with ID starting with e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.550890 4756 scope.go:117] "RemoveContainer" containerID="8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551052 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} err="failed to get container status \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": rpc error: code = NotFound desc = could not find container \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": container with ID starting with 8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551071 4756 scope.go:117] "RemoveContainer" containerID="7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551234 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} err="failed to get container status \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": rpc error: code = NotFound desc = could not find container \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": container with ID starting with 7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551253 4756 scope.go:117] "RemoveContainer" containerID="7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551401 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} err="failed to get container status \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": rpc error: code = NotFound desc = could not find container \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": container with ID starting with 7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551417 4756 scope.go:117] "RemoveContainer" containerID="8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551608 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} err="failed to get container status \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": rpc error: code = NotFound desc = could not find container \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": container with ID starting with 8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551627 4756 scope.go:117] "RemoveContainer" containerID="f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551770 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b"} err="failed to get container status \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": rpc error: code = NotFound desc = could not find container \"f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b\": container with ID starting with f2d19346471d95863cf5ca4b1256d4efd664f1a8f862f4a52cb2f385b9b9bf3b not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551787 4756 scope.go:117] "RemoveContainer" containerID="c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.551976 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f"} err="failed to get container status \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": rpc error: code = NotFound desc = could not find container \"c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f\": container with ID starting with c105813a9210b279b3e273614e5e4b43d8b8895d807350ded4619bfde41f4d1f not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552004 4756 scope.go:117] "RemoveContainer" containerID="1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552175 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815"} err="failed to get container status \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": rpc error: code = NotFound desc = could not find container \"1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815\": container with ID starting with 1eda1baf012db50692d2bdec4877fa3be38df77b682c3f15e45fadde63e68815 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552213 4756 scope.go:117] "RemoveContainer" containerID="2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552378 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a"} err="failed to get container status \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": rpc error: code = NotFound desc = could not find container \"2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a\": container with ID starting with 2fe751c99365db557defa74b2cb628f8a0e26e9da5178580a1a96d6475befd3a not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552402 4756 scope.go:117] "RemoveContainer" containerID="7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552565 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b"} err="failed to get container status \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": rpc error: code = NotFound desc = could not find container \"7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b\": container with ID starting with 7c76b0024248968c7c9eefb5b529472944b4ec5c9ccb17d2d2124ea86648026b not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552582 4756 scope.go:117] "RemoveContainer" containerID="e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552735 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90"} err="failed to get container status \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": rpc error: code = NotFound desc = could not find container \"e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90\": container with ID starting with e599d348dbfd0a8f7f61f7181c7afa10fd7ce35ac00053907f37137c3d2ebc90 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552760 4756 scope.go:117] "RemoveContainer" containerID="8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552915 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb"} err="failed to get container status \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": rpc error: code = NotFound desc = could not find container \"8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb\": container with ID starting with 8fefdf9354137c960e7dbbd169a7c50be9cafeb5331816f716fd729e80e706fb not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.552931 4756 scope.go:117] "RemoveContainer" containerID="7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.553073 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391"} err="failed to get container status \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": rpc error: code = NotFound desc = could not find container \"7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391\": container with ID starting with 7c5f0f4b1a3941428186a6741a8c60a09dfd191a6f736b15134eae97f9297391 not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.553097 4756 scope.go:117] "RemoveContainer" containerID="7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.553364 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc"} err="failed to get container status \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": rpc error: code = NotFound desc = could not find container \"7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc\": container with ID starting with 7098038138c8c0b614490a27ec0eb1eba6a109924dde72136d208737fa8295fc not found: ID does not exist" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.553407 4756 scope.go:117] "RemoveContainer" containerID="8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4" Feb 03 09:22:00 crc kubenswrapper[4756]: I0203 09:22:00.553641 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4"} err="failed to get container status \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": rpc error: code = NotFound desc = could not find container \"8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4\": container with ID starting with 8cd19e2dff1b9ec5377f540a8bddf0dfb249ca026e6526d97f42480ec3b326a4 not found: ID does not exist" Feb 03 09:22:01 crc kubenswrapper[4756]: I0203 09:22:01.355123 4756 generic.go:334] "Generic (PLEG): container finished" podID="fc550307-2c84-4d4f-ae1a-e415be21241d" containerID="988ce696d695cdf18ceabe95c7d5febc09556c12a02dfa54d247708de1bcea8f" exitCode=0 Feb 03 09:22:01 crc kubenswrapper[4756]: I0203 09:22:01.355194 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerDied","Data":"988ce696d695cdf18ceabe95c7d5febc09556c12a02dfa54d247708de1bcea8f"} Feb 03 09:22:01 crc kubenswrapper[4756]: I0203 09:22:01.620615 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705" path="/var/lib/kubelet/pods/3df6b15e-d2d2-4d0e-b7fe-e674fa2a4705/volumes" Feb 03 09:22:02 crc kubenswrapper[4756]: I0203 09:22:02.363745 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"7ace4cb188158f7a03311c0709673ba0115fb88674200e04e6a1209bcc5a9204"} Feb 03 09:22:02 crc kubenswrapper[4756]: I0203 09:22:02.363794 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"b7963866584a3a25d4c3a019ec0266cfef510a817f09f9b59f9df5eba2cbcfdf"} Feb 03 09:22:02 crc kubenswrapper[4756]: I0203 09:22:02.363804 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"6fc7b960d86ac58f2e8d4e6dec8f9495e69ff1f022826d1f7759c68c380e1c52"} Feb 03 09:22:02 crc kubenswrapper[4756]: I0203 09:22:02.363813 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"bce1f1f17029a342d70d16b4b3d9661a5eec05bdff07dc3f9197238b6fe853f4"} Feb 03 09:22:02 crc kubenswrapper[4756]: I0203 09:22:02.363823 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"ec3c7d8e060cab427148d7fb3f7195ad4ddc84da5ff81b2bf2ada3d469d9382e"} Feb 03 09:22:02 crc kubenswrapper[4756]: I0203 09:22:02.363833 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"57b6af359a25327f3e727c65f07b2290a72dcb31e61ef504afe57e004e705432"} Feb 03 09:22:04 crc kubenswrapper[4756]: I0203 09:22:04.381401 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"ffe4e6e701d35a5669b3749a25eaf46ff9cf421533ef81b0be93e5e955bf3d64"} Feb 03 09:22:07 crc kubenswrapper[4756]: I0203 09:22:07.402512 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" event={"ID":"fc550307-2c84-4d4f-ae1a-e415be21241d","Type":"ContainerStarted","Data":"7deaa9fa64e7fd8c81f9f9cfd8f0f59432209130ab137ab2ee727e3d35c2d01f"} Feb 03 09:22:07 crc kubenswrapper[4756]: I0203 09:22:07.402930 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:07 crc kubenswrapper[4756]: I0203 09:22:07.402951 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:07 crc kubenswrapper[4756]: I0203 09:22:07.402962 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:07 crc kubenswrapper[4756]: I0203 09:22:07.431053 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:07 crc kubenswrapper[4756]: I0203 09:22:07.432514 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:07 crc kubenswrapper[4756]: I0203 09:22:07.465919 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" podStartSLOduration=8.465899492 podStartE2EDuration="8.465899492s" podCreationTimestamp="2026-02-03 09:21:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:22:07.438596261 +0000 UTC m=+718.589063656" watchObservedRunningTime="2026-02-03 09:22:07.465899492 +0000 UTC m=+718.616366867" Feb 03 09:22:09 crc kubenswrapper[4756]: I0203 09:22:09.933428 4756 scope.go:117] "RemoveContainer" containerID="d08dba08ec46e082c5086d9c0eeeb4adebce763af1ec5e3c03edcf4208f5355b" Feb 03 09:22:10 crc kubenswrapper[4756]: I0203 09:22:10.418530 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/2.log" Feb 03 09:22:11 crc kubenswrapper[4756]: I0203 09:22:11.614000 4756 scope.go:117] "RemoveContainer" containerID="b0a499eb23a8e69d38e737205e6ed33a99d4a789663122520887ad03b0c3fb57" Feb 03 09:22:11 crc kubenswrapper[4756]: E0203 09:22:11.615154 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6srkq_openshift-multus(89e5d60f-f690-486f-b078-5ae9f98e1f3a)\"" pod="openshift-multus/multus-6srkq" podUID="89e5d60f-f690-486f-b078-5ae9f98e1f3a" Feb 03 09:22:13 crc kubenswrapper[4756]: I0203 09:22:13.566506 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:22:13 crc kubenswrapper[4756]: I0203 09:22:13.566585 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:22:25 crc kubenswrapper[4756]: I0203 09:22:25.614618 4756 scope.go:117] "RemoveContainer" containerID="b0a499eb23a8e69d38e737205e6ed33a99d4a789663122520887ad03b0c3fb57" Feb 03 09:22:26 crc kubenswrapper[4756]: I0203 09:22:26.507743 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6srkq_89e5d60f-f690-486f-b078-5ae9f98e1f3a/kube-multus/2.log" Feb 03 09:22:26 crc kubenswrapper[4756]: I0203 09:22:26.508060 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6srkq" event={"ID":"89e5d60f-f690-486f-b078-5ae9f98e1f3a","Type":"ContainerStarted","Data":"dab261e4b121744fd863cf0c4594b4f5e184e0d9ff8668ae74dc297482bf5db8"} Feb 03 09:22:30 crc kubenswrapper[4756]: I0203 09:22:30.258226 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vlrv6" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.429687 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7"] Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.431113 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.436798 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7"] Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.437255 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.482428 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.482492 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.482626 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlct6\" (UniqueName: \"kubernetes.io/projected/5db72b48-6cac-4211-a692-0e5d98e6d9e8-kube-api-access-nlct6\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.583433 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlct6\" (UniqueName: \"kubernetes.io/projected/5db72b48-6cac-4211-a692-0e5d98e6d9e8-kube-api-access-nlct6\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.583616 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.583648 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.584165 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.584195 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.611500 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlct6\" (UniqueName: \"kubernetes.io/projected/5db72b48-6cac-4211-a692-0e5d98e6d9e8-kube-api-access-nlct6\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.746145 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:31 crc kubenswrapper[4756]: I0203 09:22:31.963630 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7"] Feb 03 09:22:31 crc kubenswrapper[4756]: W0203 09:22:31.968477 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5db72b48_6cac_4211_a692_0e5d98e6d9e8.slice/crio-a03dca70cba1c3c9cffb178632aa002d1cd86c3586f994b5fe597618a2101c2c WatchSource:0}: Error finding container a03dca70cba1c3c9cffb178632aa002d1cd86c3586f994b5fe597618a2101c2c: Status 404 returned error can't find the container with id a03dca70cba1c3c9cffb178632aa002d1cd86c3586f994b5fe597618a2101c2c Feb 03 09:22:32 crc kubenswrapper[4756]: I0203 09:22:32.541294 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" event={"ID":"5db72b48-6cac-4211-a692-0e5d98e6d9e8","Type":"ContainerStarted","Data":"b4693e2d4ca4edba4863d4ec97cd6219aa4855caf53402c092bcdad069c50cfe"} Feb 03 09:22:32 crc kubenswrapper[4756]: I0203 09:22:32.541702 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" event={"ID":"5db72b48-6cac-4211-a692-0e5d98e6d9e8","Type":"ContainerStarted","Data":"a03dca70cba1c3c9cffb178632aa002d1cd86c3586f994b5fe597618a2101c2c"} Feb 03 09:22:33 crc kubenswrapper[4756]: I0203 09:22:33.546965 4756 generic.go:334] "Generic (PLEG): container finished" podID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerID="b4693e2d4ca4edba4863d4ec97cd6219aa4855caf53402c092bcdad069c50cfe" exitCode=0 Feb 03 09:22:33 crc kubenswrapper[4756]: I0203 09:22:33.547056 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" event={"ID":"5db72b48-6cac-4211-a692-0e5d98e6d9e8","Type":"ContainerDied","Data":"b4693e2d4ca4edba4863d4ec97cd6219aa4855caf53402c092bcdad069c50cfe"} Feb 03 09:22:35 crc kubenswrapper[4756]: I0203 09:22:35.558522 4756 generic.go:334] "Generic (PLEG): container finished" podID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerID="9b85b24d8cf3fc04b72ddc4b65158ae6b6429ff6f536f061926531beb3db6a26" exitCode=0 Feb 03 09:22:35 crc kubenswrapper[4756]: I0203 09:22:35.558570 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" event={"ID":"5db72b48-6cac-4211-a692-0e5d98e6d9e8","Type":"ContainerDied","Data":"9b85b24d8cf3fc04b72ddc4b65158ae6b6429ff6f536f061926531beb3db6a26"} Feb 03 09:22:36 crc kubenswrapper[4756]: I0203 09:22:36.567503 4756 generic.go:334] "Generic (PLEG): container finished" podID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerID="9a137000cb1701161478bd848c1995bdf9775a5937261cce3708b677bde34421" exitCode=0 Feb 03 09:22:36 crc kubenswrapper[4756]: I0203 09:22:36.567562 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" event={"ID":"5db72b48-6cac-4211-a692-0e5d98e6d9e8","Type":"ContainerDied","Data":"9a137000cb1701161478bd848c1995bdf9775a5937261cce3708b677bde34421"} Feb 03 09:22:37 crc kubenswrapper[4756]: I0203 09:22:37.817599 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:37 crc kubenswrapper[4756]: I0203 09:22:37.959327 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-bundle" (OuterVolumeSpecName: "bundle") pod "5db72b48-6cac-4211-a692-0e5d98e6d9e8" (UID: "5db72b48-6cac-4211-a692-0e5d98e6d9e8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:22:37 crc kubenswrapper[4756]: I0203 09:22:37.958434 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-bundle\") pod \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " Feb 03 09:22:37 crc kubenswrapper[4756]: I0203 09:22:37.959419 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlct6\" (UniqueName: \"kubernetes.io/projected/5db72b48-6cac-4211-a692-0e5d98e6d9e8-kube-api-access-nlct6\") pod \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " Feb 03 09:22:37 crc kubenswrapper[4756]: I0203 09:22:37.960301 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-util\") pod \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\" (UID: \"5db72b48-6cac-4211-a692-0e5d98e6d9e8\") " Feb 03 09:22:37 crc kubenswrapper[4756]: I0203 09:22:37.960507 4756 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:22:37 crc kubenswrapper[4756]: I0203 09:22:37.965959 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5db72b48-6cac-4211-a692-0e5d98e6d9e8-kube-api-access-nlct6" (OuterVolumeSpecName: "kube-api-access-nlct6") pod "5db72b48-6cac-4211-a692-0e5d98e6d9e8" (UID: "5db72b48-6cac-4211-a692-0e5d98e6d9e8"). InnerVolumeSpecName "kube-api-access-nlct6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:22:37 crc kubenswrapper[4756]: I0203 09:22:37.971389 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-util" (OuterVolumeSpecName: "util") pod "5db72b48-6cac-4211-a692-0e5d98e6d9e8" (UID: "5db72b48-6cac-4211-a692-0e5d98e6d9e8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:22:38 crc kubenswrapper[4756]: I0203 09:22:38.063719 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlct6\" (UniqueName: \"kubernetes.io/projected/5db72b48-6cac-4211-a692-0e5d98e6d9e8-kube-api-access-nlct6\") on node \"crc\" DevicePath \"\"" Feb 03 09:22:38 crc kubenswrapper[4756]: I0203 09:22:38.063811 4756 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5db72b48-6cac-4211-a692-0e5d98e6d9e8-util\") on node \"crc\" DevicePath \"\"" Feb 03 09:22:38 crc kubenswrapper[4756]: I0203 09:22:38.582603 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" event={"ID":"5db72b48-6cac-4211-a692-0e5d98e6d9e8","Type":"ContainerDied","Data":"a03dca70cba1c3c9cffb178632aa002d1cd86c3586f994b5fe597618a2101c2c"} Feb 03 09:22:38 crc kubenswrapper[4756]: I0203 09:22:38.582652 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a03dca70cba1c3c9cffb178632aa002d1cd86c3586f994b5fe597618a2101c2c" Feb 03 09:22:38 crc kubenswrapper[4756]: I0203 09:22:38.582706 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.958985 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-wvckj"] Feb 03 09:22:39 crc kubenswrapper[4756]: E0203 09:22:39.959189 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerName="extract" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.959200 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerName="extract" Feb 03 09:22:39 crc kubenswrapper[4756]: E0203 09:22:39.959215 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerName="pull" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.959222 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerName="pull" Feb 03 09:22:39 crc kubenswrapper[4756]: E0203 09:22:39.959234 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerName="util" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.959239 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerName="util" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.959337 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5db72b48-6cac-4211-a692-0e5d98e6d9e8" containerName="extract" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.959770 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-wvckj" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.962389 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.963675 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.964067 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-v9g9r" Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.976726 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-wvckj"] Feb 03 09:22:39 crc kubenswrapper[4756]: I0203 09:22:39.990663 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzrm8\" (UniqueName: \"kubernetes.io/projected/6374dff8-08c2-49fe-b5f4-0eb5142f61bf-kube-api-access-dzrm8\") pod \"nmstate-operator-646758c888-wvckj\" (UID: \"6374dff8-08c2-49fe-b5f4-0eb5142f61bf\") " pod="openshift-nmstate/nmstate-operator-646758c888-wvckj" Feb 03 09:22:40 crc kubenswrapper[4756]: I0203 09:22:40.092483 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzrm8\" (UniqueName: \"kubernetes.io/projected/6374dff8-08c2-49fe-b5f4-0eb5142f61bf-kube-api-access-dzrm8\") pod \"nmstate-operator-646758c888-wvckj\" (UID: \"6374dff8-08c2-49fe-b5f4-0eb5142f61bf\") " pod="openshift-nmstate/nmstate-operator-646758c888-wvckj" Feb 03 09:22:40 crc kubenswrapper[4756]: I0203 09:22:40.111206 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzrm8\" (UniqueName: \"kubernetes.io/projected/6374dff8-08c2-49fe-b5f4-0eb5142f61bf-kube-api-access-dzrm8\") pod \"nmstate-operator-646758c888-wvckj\" (UID: \"6374dff8-08c2-49fe-b5f4-0eb5142f61bf\") " pod="openshift-nmstate/nmstate-operator-646758c888-wvckj" Feb 03 09:22:40 crc kubenswrapper[4756]: I0203 09:22:40.274965 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-wvckj" Feb 03 09:22:40 crc kubenswrapper[4756]: I0203 09:22:40.485974 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-wvckj"] Feb 03 09:22:40 crc kubenswrapper[4756]: W0203 09:22:40.501521 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6374dff8_08c2_49fe_b5f4_0eb5142f61bf.slice/crio-1fd200420aeff7dfdfe31d2f51b56534955a3e73cbdd6bbbdcbedac554e6bb2f WatchSource:0}: Error finding container 1fd200420aeff7dfdfe31d2f51b56534955a3e73cbdd6bbbdcbedac554e6bb2f: Status 404 returned error can't find the container with id 1fd200420aeff7dfdfe31d2f51b56534955a3e73cbdd6bbbdcbedac554e6bb2f Feb 03 09:22:40 crc kubenswrapper[4756]: I0203 09:22:40.595067 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-wvckj" event={"ID":"6374dff8-08c2-49fe-b5f4-0eb5142f61bf","Type":"ContainerStarted","Data":"1fd200420aeff7dfdfe31d2f51b56534955a3e73cbdd6bbbdcbedac554e6bb2f"} Feb 03 09:22:42 crc kubenswrapper[4756]: I0203 09:22:42.370340 4756 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 03 09:22:42 crc kubenswrapper[4756]: I0203 09:22:42.608147 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-wvckj" event={"ID":"6374dff8-08c2-49fe-b5f4-0eb5142f61bf","Type":"ContainerStarted","Data":"9f2e55e354b851d8419ab5ac717c71df7e0aae4d10349b825887b8232835d6cf"} Feb 03 09:22:42 crc kubenswrapper[4756]: I0203 09:22:42.634782 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-wvckj" podStartSLOduration=1.9542252549999999 podStartE2EDuration="3.63475909s" podCreationTimestamp="2026-02-03 09:22:39 +0000 UTC" firstStartedPulling="2026-02-03 09:22:40.504795412 +0000 UTC m=+751.655262787" lastFinishedPulling="2026-02-03 09:22:42.185329237 +0000 UTC m=+753.335796622" observedRunningTime="2026-02-03 09:22:42.621038923 +0000 UTC m=+753.771506318" watchObservedRunningTime="2026-02-03 09:22:42.63475909 +0000 UTC m=+753.785226465" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.537777 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.539152 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.541751 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.542476 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-2z97n" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.555030 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.563178 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-75x2d"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.566225 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.566297 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.572309 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.572429 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.573279 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c20ecab788892a342dab2838ffafa76a4cbfbef0fe19794f920df87b66dcf0b1"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.573333 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://c20ecab788892a342dab2838ffafa76a4cbfbef0fe19794f920df87b66dcf0b1" gracePeriod=600 Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.588511 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-75x2d"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.638552 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-fv282"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.639264 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.639378 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r94cr\" (UniqueName: \"kubernetes.io/projected/b9dfc50f-d7ab-4806-aac1-70e36788c60e-kube-api-access-r94cr\") pod \"nmstate-webhook-8474b5b9d8-cjp9b\" (UID: \"b9dfc50f-d7ab-4806-aac1-70e36788c60e\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.639428 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7gvc\" (UniqueName: \"kubernetes.io/projected/63b70a58-8d39-418a-bedc-93f9a94540a4-kube-api-access-q7gvc\") pod \"nmstate-metrics-54757c584b-75x2d\" (UID: \"63b70a58-8d39-418a-bedc-93f9a94540a4\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.639475 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b9dfc50f-d7ab-4806-aac1-70e36788c60e-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-cjp9b\" (UID: \"b9dfc50f-d7ab-4806-aac1-70e36788c60e\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.715727 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.716336 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.719734 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.720041 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-5sqft" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.720361 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.739994 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac470c2b-de36-4daa-bce0-8163e27e43e7-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740035 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7gvc\" (UniqueName: \"kubernetes.io/projected/63b70a58-8d39-418a-bedc-93f9a94540a4-kube-api-access-q7gvc\") pod \"nmstate-metrics-54757c584b-75x2d\" (UID: \"63b70a58-8d39-418a-bedc-93f9a94540a4\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740057 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-nmstate-lock\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740079 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-dbus-socket\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740102 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b9dfc50f-d7ab-4806-aac1-70e36788c60e-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-cjp9b\" (UID: \"b9dfc50f-d7ab-4806-aac1-70e36788c60e\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740118 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-ovs-socket\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740144 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rz22\" (UniqueName: \"kubernetes.io/projected/ac470c2b-de36-4daa-bce0-8163e27e43e7-kube-api-access-2rz22\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740171 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ac470c2b-de36-4daa-bce0-8163e27e43e7-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740195 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r94cr\" (UniqueName: \"kubernetes.io/projected/b9dfc50f-d7ab-4806-aac1-70e36788c60e-kube-api-access-r94cr\") pod \"nmstate-webhook-8474b5b9d8-cjp9b\" (UID: \"b9dfc50f-d7ab-4806-aac1-70e36788c60e\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.740231 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k68qc\" (UniqueName: \"kubernetes.io/projected/e827e9ff-5a84-42cc-9927-56904973c3ad-kube-api-access-k68qc\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: E0203 09:22:43.740662 4756 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Feb 03 09:22:43 crc kubenswrapper[4756]: E0203 09:22:43.740737 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9dfc50f-d7ab-4806-aac1-70e36788c60e-tls-key-pair podName:b9dfc50f-d7ab-4806-aac1-70e36788c60e nodeName:}" failed. No retries permitted until 2026-02-03 09:22:44.240714675 +0000 UTC m=+755.391182060 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/b9dfc50f-d7ab-4806-aac1-70e36788c60e-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-cjp9b" (UID: "b9dfc50f-d7ab-4806-aac1-70e36788c60e") : secret "openshift-nmstate-webhook" not found Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.742654 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.763954 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7gvc\" (UniqueName: \"kubernetes.io/projected/63b70a58-8d39-418a-bedc-93f9a94540a4-kube-api-access-q7gvc\") pod \"nmstate-metrics-54757c584b-75x2d\" (UID: \"63b70a58-8d39-418a-bedc-93f9a94540a4\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.764909 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r94cr\" (UniqueName: \"kubernetes.io/projected/b9dfc50f-d7ab-4806-aac1-70e36788c60e-kube-api-access-r94cr\") pod \"nmstate-webhook-8474b5b9d8-cjp9b\" (UID: \"b9dfc50f-d7ab-4806-aac1-70e36788c60e\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841269 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-dbus-socket\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841501 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-ovs-socket\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841598 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rz22\" (UniqueName: \"kubernetes.io/projected/ac470c2b-de36-4daa-bce0-8163e27e43e7-kube-api-access-2rz22\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841585 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-ovs-socket\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841677 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-dbus-socket\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841694 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ac470c2b-de36-4daa-bce0-8163e27e43e7-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841781 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k68qc\" (UniqueName: \"kubernetes.io/projected/e827e9ff-5a84-42cc-9927-56904973c3ad-kube-api-access-k68qc\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841834 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac470c2b-de36-4daa-bce0-8163e27e43e7-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841872 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-nmstate-lock\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.841960 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e827e9ff-5a84-42cc-9927-56904973c3ad-nmstate-lock\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: E0203 09:22:43.842078 4756 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 03 09:22:43 crc kubenswrapper[4756]: E0203 09:22:43.842159 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac470c2b-de36-4daa-bce0-8163e27e43e7-plugin-serving-cert podName:ac470c2b-de36-4daa-bce0-8163e27e43e7 nodeName:}" failed. No retries permitted until 2026-02-03 09:22:44.342137263 +0000 UTC m=+755.492604638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/ac470c2b-de36-4daa-bce0-8163e27e43e7-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-zscxn" (UID: "ac470c2b-de36-4daa-bce0-8163e27e43e7") : secret "plugin-serving-cert" not found Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.842667 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ac470c2b-de36-4daa-bce0-8163e27e43e7-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.862895 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k68qc\" (UniqueName: \"kubernetes.io/projected/e827e9ff-5a84-42cc-9927-56904973c3ad-kube-api-access-k68qc\") pod \"nmstate-handler-fv282\" (UID: \"e827e9ff-5a84-42cc-9927-56904973c3ad\") " pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.872074 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rz22\" (UniqueName: \"kubernetes.io/projected/ac470c2b-de36-4daa-bce0-8163e27e43e7-kube-api-access-2rz22\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.908866 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.939733 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-97df9c74f-zxlns"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.940628 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.958736 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-97df9c74f-zxlns"] Feb 03 09:22:43 crc kubenswrapper[4756]: I0203 09:22:43.966473 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:43 crc kubenswrapper[4756]: W0203 09:22:43.989258 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode827e9ff_5a84_42cc_9927_56904973c3ad.slice/crio-3e15750c7a1bd518ad91aa4dc8059ab8021c283ea157eca9dfdb0c97abc63f73 WatchSource:0}: Error finding container 3e15750c7a1bd518ad91aa4dc8059ab8021c283ea157eca9dfdb0c97abc63f73: Status 404 returned error can't find the container with id 3e15750c7a1bd518ad91aa4dc8059ab8021c283ea157eca9dfdb0c97abc63f73 Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.046122 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-oauth-serving-cert\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.046479 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-serving-cert\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.046504 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-trusted-ca-bundle\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.046540 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-service-ca\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.046561 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-oauth-config\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.046717 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q875l\" (UniqueName: \"kubernetes.io/projected/b2da71aa-f83e-4a70-9dae-2e40d5606874-kube-api-access-q875l\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.046768 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-config\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.137633 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-75x2d"] Feb 03 09:22:44 crc kubenswrapper[4756]: W0203 09:22:44.144522 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b70a58_8d39_418a_bedc_93f9a94540a4.slice/crio-e38fb83f36b3f28c0e3d6278aaa4a31d59cf4b33b875f1e279f9dcf19d5b7a5e WatchSource:0}: Error finding container e38fb83f36b3f28c0e3d6278aaa4a31d59cf4b33b875f1e279f9dcf19d5b7a5e: Status 404 returned error can't find the container with id e38fb83f36b3f28c0e3d6278aaa4a31d59cf4b33b875f1e279f9dcf19d5b7a5e Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.147416 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-serving-cert\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.147448 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-trusted-ca-bundle\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.147488 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-service-ca\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.147509 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-oauth-config\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.147548 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q875l\" (UniqueName: \"kubernetes.io/projected/b2da71aa-f83e-4a70-9dae-2e40d5606874-kube-api-access-q875l\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.147575 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-config\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.147604 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-oauth-serving-cert\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.148633 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-service-ca\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.148650 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-oauth-serving-cert\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.148720 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-config\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.149113 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2da71aa-f83e-4a70-9dae-2e40d5606874-trusted-ca-bundle\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.156428 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-oauth-config\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.157246 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b2da71aa-f83e-4a70-9dae-2e40d5606874-console-serving-cert\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.166199 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q875l\" (UniqueName: \"kubernetes.io/projected/b2da71aa-f83e-4a70-9dae-2e40d5606874-kube-api-access-q875l\") pod \"console-97df9c74f-zxlns\" (UID: \"b2da71aa-f83e-4a70-9dae-2e40d5606874\") " pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.248475 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b9dfc50f-d7ab-4806-aac1-70e36788c60e-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-cjp9b\" (UID: \"b9dfc50f-d7ab-4806-aac1-70e36788c60e\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.252231 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b9dfc50f-d7ab-4806-aac1-70e36788c60e-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-cjp9b\" (UID: \"b9dfc50f-d7ab-4806-aac1-70e36788c60e\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.268049 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.354313 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac470c2b-de36-4daa-bce0-8163e27e43e7-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.362048 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac470c2b-de36-4daa-bce0-8163e27e43e7-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-zscxn\" (UID: \"ac470c2b-de36-4daa-bce0-8163e27e43e7\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.455274 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.478590 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-97df9c74f-zxlns"] Feb 03 09:22:44 crc kubenswrapper[4756]: W0203 09:22:44.484478 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2da71aa_f83e_4a70_9dae_2e40d5606874.slice/crio-6746af76e42603d9a5a3b9fb1a0a4a391b9ec87452aa338039fa532002fdfe94 WatchSource:0}: Error finding container 6746af76e42603d9a5a3b9fb1a0a4a391b9ec87452aa338039fa532002fdfe94: Status 404 returned error can't find the container with id 6746af76e42603d9a5a3b9fb1a0a4a391b9ec87452aa338039fa532002fdfe94 Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.627580 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-97df9c74f-zxlns" event={"ID":"b2da71aa-f83e-4a70-9dae-2e40d5606874","Type":"ContainerStarted","Data":"6746af76e42603d9a5a3b9fb1a0a4a391b9ec87452aa338039fa532002fdfe94"} Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.631870 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b"] Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.637284 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" event={"ID":"63b70a58-8d39-418a-bedc-93f9a94540a4","Type":"ContainerStarted","Data":"e38fb83f36b3f28c0e3d6278aaa4a31d59cf4b33b875f1e279f9dcf19d5b7a5e"} Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.638372 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fv282" event={"ID":"e827e9ff-5a84-42cc-9927-56904973c3ad","Type":"ContainerStarted","Data":"3e15750c7a1bd518ad91aa4dc8059ab8021c283ea157eca9dfdb0c97abc63f73"} Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.641112 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="c20ecab788892a342dab2838ffafa76a4cbfbef0fe19794f920df87b66dcf0b1" exitCode=0 Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.641147 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"c20ecab788892a342dab2838ffafa76a4cbfbef0fe19794f920df87b66dcf0b1"} Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.641197 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"ef696d62114310b595359fdf0f823a6781a2df097280882ce3528c732940055c"} Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.641217 4756 scope.go:117] "RemoveContainer" containerID="47fef4302ae2be888f70b5160d00aea824e230accaf11fd9a2599e23eb41dc66" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.643008 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" Feb 03 09:22:44 crc kubenswrapper[4756]: I0203 09:22:44.835201 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn"] Feb 03 09:22:44 crc kubenswrapper[4756]: W0203 09:22:44.843002 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac470c2b_de36_4daa_bce0_8163e27e43e7.slice/crio-ac67e06f3315d1145e7b639f0acd68bc075e3159296fd8a91c81e8410af69232 WatchSource:0}: Error finding container ac67e06f3315d1145e7b639f0acd68bc075e3159296fd8a91c81e8410af69232: Status 404 returned error can't find the container with id ac67e06f3315d1145e7b639f0acd68bc075e3159296fd8a91c81e8410af69232 Feb 03 09:22:45 crc kubenswrapper[4756]: I0203 09:22:45.649207 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-97df9c74f-zxlns" event={"ID":"b2da71aa-f83e-4a70-9dae-2e40d5606874","Type":"ContainerStarted","Data":"9f9efabd9fa082a74911c3dd205a9a175e41724636d3289f46e52c873b197a7e"} Feb 03 09:22:45 crc kubenswrapper[4756]: I0203 09:22:45.654332 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" event={"ID":"b9dfc50f-d7ab-4806-aac1-70e36788c60e","Type":"ContainerStarted","Data":"49a0084a9a2515c52e97adeba11510b0347d22995a222afabb493ae4e9d97a7e"} Feb 03 09:22:45 crc kubenswrapper[4756]: I0203 09:22:45.655556 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" event={"ID":"ac470c2b-de36-4daa-bce0-8163e27e43e7","Type":"ContainerStarted","Data":"ac67e06f3315d1145e7b639f0acd68bc075e3159296fd8a91c81e8410af69232"} Feb 03 09:22:45 crc kubenswrapper[4756]: I0203 09:22:45.666539 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-97df9c74f-zxlns" podStartSLOduration=2.666521177 podStartE2EDuration="2.666521177s" podCreationTimestamp="2026-02-03 09:22:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:22:45.665403963 +0000 UTC m=+756.815871348" watchObservedRunningTime="2026-02-03 09:22:45.666521177 +0000 UTC m=+756.816988552" Feb 03 09:22:46 crc kubenswrapper[4756]: I0203 09:22:46.665282 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" event={"ID":"63b70a58-8d39-418a-bedc-93f9a94540a4","Type":"ContainerStarted","Data":"969fc818341dd0301a94bb12389b26361b4ae2786f4379207f4197469332370a"} Feb 03 09:22:46 crc kubenswrapper[4756]: I0203 09:22:46.668622 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fv282" event={"ID":"e827e9ff-5a84-42cc-9927-56904973c3ad","Type":"ContainerStarted","Data":"d96dabb341c1caad731ba03edf62871a72e06a416073859d5ff88efaf13dd364"} Feb 03 09:22:46 crc kubenswrapper[4756]: I0203 09:22:46.669534 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:46 crc kubenswrapper[4756]: I0203 09:22:46.672968 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" event={"ID":"b9dfc50f-d7ab-4806-aac1-70e36788c60e","Type":"ContainerStarted","Data":"259e4fbb654bef7ffd2622720ce6af0627b4b4ec77be1ca6693dfaf81191dd8a"} Feb 03 09:22:46 crc kubenswrapper[4756]: I0203 09:22:46.673005 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:22:46 crc kubenswrapper[4756]: I0203 09:22:46.691517 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-fv282" podStartSLOduration=1.479733567 podStartE2EDuration="3.691499921s" podCreationTimestamp="2026-02-03 09:22:43 +0000 UTC" firstStartedPulling="2026-02-03 09:22:43.992156565 +0000 UTC m=+755.142623940" lastFinishedPulling="2026-02-03 09:22:46.203922909 +0000 UTC m=+757.354390294" observedRunningTime="2026-02-03 09:22:46.688761255 +0000 UTC m=+757.839228640" watchObservedRunningTime="2026-02-03 09:22:46.691499921 +0000 UTC m=+757.841967296" Feb 03 09:22:46 crc kubenswrapper[4756]: I0203 09:22:46.708366 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" podStartSLOduration=2.132451988 podStartE2EDuration="3.708340055s" podCreationTimestamp="2026-02-03 09:22:43 +0000 UTC" firstStartedPulling="2026-02-03 09:22:44.652601348 +0000 UTC m=+755.803068723" lastFinishedPulling="2026-02-03 09:22:46.228489405 +0000 UTC m=+757.378956790" observedRunningTime="2026-02-03 09:22:46.708219111 +0000 UTC m=+757.858686486" watchObservedRunningTime="2026-02-03 09:22:46.708340055 +0000 UTC m=+757.858807430" Feb 03 09:22:47 crc kubenswrapper[4756]: I0203 09:22:47.680698 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" event={"ID":"ac470c2b-de36-4daa-bce0-8163e27e43e7","Type":"ContainerStarted","Data":"658a6510331094c31137b0c1aa99a2b0af90925baf6c7c1306e2c25d2e20548b"} Feb 03 09:22:47 crc kubenswrapper[4756]: I0203 09:22:47.696761 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-zscxn" podStartSLOduration=2.419752304 podStartE2EDuration="4.69674063s" podCreationTimestamp="2026-02-03 09:22:43 +0000 UTC" firstStartedPulling="2026-02-03 09:22:44.84633555 +0000 UTC m=+755.996802925" lastFinishedPulling="2026-02-03 09:22:47.123323856 +0000 UTC m=+758.273791251" observedRunningTime="2026-02-03 09:22:47.694881642 +0000 UTC m=+758.845349037" watchObservedRunningTime="2026-02-03 09:22:47.69674063 +0000 UTC m=+758.847207995" Feb 03 09:22:48 crc kubenswrapper[4756]: I0203 09:22:48.688549 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" event={"ID":"63b70a58-8d39-418a-bedc-93f9a94540a4","Type":"ContainerStarted","Data":"89d7996be6ee14f1251e3905428b31f8ee60ddb00e0da378aae098b4873142d7"} Feb 03 09:22:48 crc kubenswrapper[4756]: I0203 09:22:48.722523 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-75x2d" podStartSLOduration=1.335384082 podStartE2EDuration="5.722497428s" podCreationTimestamp="2026-02-03 09:22:43 +0000 UTC" firstStartedPulling="2026-02-03 09:22:44.147515032 +0000 UTC m=+755.297982407" lastFinishedPulling="2026-02-03 09:22:48.534628378 +0000 UTC m=+759.685095753" observedRunningTime="2026-02-03 09:22:48.702428413 +0000 UTC m=+759.852895788" watchObservedRunningTime="2026-02-03 09:22:48.722497428 +0000 UTC m=+759.872964803" Feb 03 09:22:53 crc kubenswrapper[4756]: I0203 09:22:53.994695 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-fv282" Feb 03 09:22:54 crc kubenswrapper[4756]: I0203 09:22:54.269658 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:54 crc kubenswrapper[4756]: I0203 09:22:54.269742 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:54 crc kubenswrapper[4756]: I0203 09:22:54.274289 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:54 crc kubenswrapper[4756]: I0203 09:22:54.729842 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-97df9c74f-zxlns" Feb 03 09:22:54 crc kubenswrapper[4756]: I0203 09:22:54.780476 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6qrwb"] Feb 03 09:23:04 crc kubenswrapper[4756]: I0203 09:23:04.463551 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cjp9b" Feb 03 09:23:15 crc kubenswrapper[4756]: I0203 09:23:15.982904 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q"] Feb 03 09:23:15 crc kubenswrapper[4756]: I0203 09:23:15.984320 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:15 crc kubenswrapper[4756]: I0203 09:23:15.989972 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 03 09:23:15 crc kubenswrapper[4756]: I0203 09:23:15.997520 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q"] Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.160052 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th6l8\" (UniqueName: \"kubernetes.io/projected/3dd53027-4438-4022-9ee5-8ed46a8537e3-kube-api-access-th6l8\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.160159 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.160228 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.261310 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th6l8\" (UniqueName: \"kubernetes.io/projected/3dd53027-4438-4022-9ee5-8ed46a8537e3-kube-api-access-th6l8\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.261358 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.261577 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.262091 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.262275 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.293923 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th6l8\" (UniqueName: \"kubernetes.io/projected/3dd53027-4438-4022-9ee5-8ed46a8537e3-kube-api-access-th6l8\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.298871 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.491487 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q"] Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.852311 4756 generic.go:334] "Generic (PLEG): container finished" podID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerID="e1cdebd35665e2f15697989c4be5f5719f18fe06971dd70545c0b86d99e60a4e" exitCode=0 Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.852375 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" event={"ID":"3dd53027-4438-4022-9ee5-8ed46a8537e3","Type":"ContainerDied","Data":"e1cdebd35665e2f15697989c4be5f5719f18fe06971dd70545c0b86d99e60a4e"} Feb 03 09:23:16 crc kubenswrapper[4756]: I0203 09:23:16.852641 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" event={"ID":"3dd53027-4438-4022-9ee5-8ed46a8537e3","Type":"ContainerStarted","Data":"a561340a77489f4c5d35158b82c3a3c536302bdfb6e9520da586e2c2c89bb552"} Feb 03 09:23:17 crc kubenswrapper[4756]: I0203 09:23:17.865139 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" event={"ID":"3dd53027-4438-4022-9ee5-8ed46a8537e3","Type":"ContainerStarted","Data":"1c5245e0863cc3e05f0295bb48324783dc74f19299611114d7cd79efc162af82"} Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.230360 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2cmfr"] Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.231937 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.255251 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2cmfr"] Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.296422 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-catalog-content\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.297490 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-utilities\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.297578 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lfnz\" (UniqueName: \"kubernetes.io/projected/e4384d13-ef52-4709-b694-8ed1628af525-kube-api-access-9lfnz\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.398540 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-catalog-content\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.398622 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-utilities\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.398650 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lfnz\" (UniqueName: \"kubernetes.io/projected/e4384d13-ef52-4709-b694-8ed1628af525-kube-api-access-9lfnz\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.399109 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-utilities\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.399120 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-catalog-content\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.418393 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lfnz\" (UniqueName: \"kubernetes.io/projected/e4384d13-ef52-4709-b694-8ed1628af525-kube-api-access-9lfnz\") pod \"redhat-operators-2cmfr\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.564904 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.874750 4756 generic.go:334] "Generic (PLEG): container finished" podID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerID="1c5245e0863cc3e05f0295bb48324783dc74f19299611114d7cd79efc162af82" exitCode=0 Feb 03 09:23:18 crc kubenswrapper[4756]: I0203 09:23:18.874825 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" event={"ID":"3dd53027-4438-4022-9ee5-8ed46a8537e3","Type":"ContainerDied","Data":"1c5245e0863cc3e05f0295bb48324783dc74f19299611114d7cd79efc162af82"} Feb 03 09:23:19 crc kubenswrapper[4756]: I0203 09:23:19.014786 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2cmfr"] Feb 03 09:23:19 crc kubenswrapper[4756]: W0203 09:23:19.020511 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4384d13_ef52_4709_b694_8ed1628af525.slice/crio-914c33ded8cd990e409ba346ad699cae014626fd64e69ab4cd7c3bd375854a99 WatchSource:0}: Error finding container 914c33ded8cd990e409ba346ad699cae014626fd64e69ab4cd7c3bd375854a99: Status 404 returned error can't find the container with id 914c33ded8cd990e409ba346ad699cae014626fd64e69ab4cd7c3bd375854a99 Feb 03 09:23:19 crc kubenswrapper[4756]: I0203 09:23:19.821431 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-6qrwb" podUID="4dad28d0-454d-434b-b111-137c215aea4a" containerName="console" containerID="cri-o://1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221" gracePeriod=15 Feb 03 09:23:19 crc kubenswrapper[4756]: I0203 09:23:19.882234 4756 generic.go:334] "Generic (PLEG): container finished" podID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerID="e443ab79752e5c33972b64d0cf724ea92c3a6321adb91e9f1681873994c078cc" exitCode=0 Feb 03 09:23:19 crc kubenswrapper[4756]: I0203 09:23:19.882362 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" event={"ID":"3dd53027-4438-4022-9ee5-8ed46a8537e3","Type":"ContainerDied","Data":"e443ab79752e5c33972b64d0cf724ea92c3a6321adb91e9f1681873994c078cc"} Feb 03 09:23:19 crc kubenswrapper[4756]: I0203 09:23:19.884766 4756 generic.go:334] "Generic (PLEG): container finished" podID="e4384d13-ef52-4709-b694-8ed1628af525" containerID="0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575" exitCode=0 Feb 03 09:23:19 crc kubenswrapper[4756]: I0203 09:23:19.884809 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cmfr" event={"ID":"e4384d13-ef52-4709-b694-8ed1628af525","Type":"ContainerDied","Data":"0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575"} Feb 03 09:23:19 crc kubenswrapper[4756]: I0203 09:23:19.884837 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cmfr" event={"ID":"e4384d13-ef52-4709-b694-8ed1628af525","Type":"ContainerStarted","Data":"914c33ded8cd990e409ba346ad699cae014626fd64e69ab4cd7c3bd375854a99"} Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.153309 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6qrwb_4dad28d0-454d-434b-b111-137c215aea4a/console/0.log" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.153379 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.318642 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-console-config\") pod \"4dad28d0-454d-434b-b111-137c215aea4a\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.318713 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-serving-cert\") pod \"4dad28d0-454d-434b-b111-137c215aea4a\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.318735 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-service-ca\") pod \"4dad28d0-454d-434b-b111-137c215aea4a\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.318882 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-oauth-serving-cert\") pod \"4dad28d0-454d-434b-b111-137c215aea4a\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.318900 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-oauth-config\") pod \"4dad28d0-454d-434b-b111-137c215aea4a\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.318917 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-trusted-ca-bundle\") pod \"4dad28d0-454d-434b-b111-137c215aea4a\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.318938 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4cg5\" (UniqueName: \"kubernetes.io/projected/4dad28d0-454d-434b-b111-137c215aea4a-kube-api-access-d4cg5\") pod \"4dad28d0-454d-434b-b111-137c215aea4a\" (UID: \"4dad28d0-454d-434b-b111-137c215aea4a\") " Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.319559 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-console-config" (OuterVolumeSpecName: "console-config") pod "4dad28d0-454d-434b-b111-137c215aea4a" (UID: "4dad28d0-454d-434b-b111-137c215aea4a"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.319880 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4dad28d0-454d-434b-b111-137c215aea4a" (UID: "4dad28d0-454d-434b-b111-137c215aea4a"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.320036 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4dad28d0-454d-434b-b111-137c215aea4a" (UID: "4dad28d0-454d-434b-b111-137c215aea4a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.320289 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-service-ca" (OuterVolumeSpecName: "service-ca") pod "4dad28d0-454d-434b-b111-137c215aea4a" (UID: "4dad28d0-454d-434b-b111-137c215aea4a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.324429 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dad28d0-454d-434b-b111-137c215aea4a-kube-api-access-d4cg5" (OuterVolumeSpecName: "kube-api-access-d4cg5") pod "4dad28d0-454d-434b-b111-137c215aea4a" (UID: "4dad28d0-454d-434b-b111-137c215aea4a"). InnerVolumeSpecName "kube-api-access-d4cg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.326808 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4dad28d0-454d-434b-b111-137c215aea4a" (UID: "4dad28d0-454d-434b-b111-137c215aea4a"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.327180 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4dad28d0-454d-434b-b111-137c215aea4a" (UID: "4dad28d0-454d-434b-b111-137c215aea4a"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.419860 4756 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.419897 4756 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.419907 4756 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.419916 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4cg5\" (UniqueName: \"kubernetes.io/projected/4dad28d0-454d-434b-b111-137c215aea4a-kube-api-access-d4cg5\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.419924 4756 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-console-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.419933 4756 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dad28d0-454d-434b-b111-137c215aea4a-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.419959 4756 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4dad28d0-454d-434b-b111-137c215aea4a-service-ca\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.899396 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6qrwb_4dad28d0-454d-434b-b111-137c215aea4a/console/0.log" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.899467 4756 generic.go:334] "Generic (PLEG): container finished" podID="4dad28d0-454d-434b-b111-137c215aea4a" containerID="1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221" exitCode=2 Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.899542 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6qrwb" event={"ID":"4dad28d0-454d-434b-b111-137c215aea4a","Type":"ContainerDied","Data":"1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221"} Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.899616 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6qrwb" event={"ID":"4dad28d0-454d-434b-b111-137c215aea4a","Type":"ContainerDied","Data":"1eb0e66ccf3bd9b01f57ba9ef6d90a956da14150825527898a44842048b76028"} Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.899639 4756 scope.go:117] "RemoveContainer" containerID="1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.899562 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6qrwb" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.924344 4756 scope.go:117] "RemoveContainer" containerID="1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221" Feb 03 09:23:20 crc kubenswrapper[4756]: E0203 09:23:20.924928 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221\": container with ID starting with 1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221 not found: ID does not exist" containerID="1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.924967 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221"} err="failed to get container status \"1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221\": rpc error: code = NotFound desc = could not find container \"1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221\": container with ID starting with 1eafaf308916171dbd1ad32f770db9fe8f9845ff8623f44e4ac17d85eecc5221 not found: ID does not exist" Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.932846 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6qrwb"] Feb 03 09:23:20 crc kubenswrapper[4756]: I0203 09:23:20.936230 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-6qrwb"] Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.153683 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.331112 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-util\") pod \"3dd53027-4438-4022-9ee5-8ed46a8537e3\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.331198 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-bundle\") pod \"3dd53027-4438-4022-9ee5-8ed46a8537e3\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.331226 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th6l8\" (UniqueName: \"kubernetes.io/projected/3dd53027-4438-4022-9ee5-8ed46a8537e3-kube-api-access-th6l8\") pod \"3dd53027-4438-4022-9ee5-8ed46a8537e3\" (UID: \"3dd53027-4438-4022-9ee5-8ed46a8537e3\") " Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.333003 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-bundle" (OuterVolumeSpecName: "bundle") pod "3dd53027-4438-4022-9ee5-8ed46a8537e3" (UID: "3dd53027-4438-4022-9ee5-8ed46a8537e3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.337940 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd53027-4438-4022-9ee5-8ed46a8537e3-kube-api-access-th6l8" (OuterVolumeSpecName: "kube-api-access-th6l8") pod "3dd53027-4438-4022-9ee5-8ed46a8537e3" (UID: "3dd53027-4438-4022-9ee5-8ed46a8537e3"). InnerVolumeSpecName "kube-api-access-th6l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.344313 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-util" (OuterVolumeSpecName: "util") pod "3dd53027-4438-4022-9ee5-8ed46a8537e3" (UID: "3dd53027-4438-4022-9ee5-8ed46a8537e3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.432999 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th6l8\" (UniqueName: \"kubernetes.io/projected/3dd53027-4438-4022-9ee5-8ed46a8537e3-kube-api-access-th6l8\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.433287 4756 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-util\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.433362 4756 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3dd53027-4438-4022-9ee5-8ed46a8537e3-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.621633 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dad28d0-454d-434b-b111-137c215aea4a" path="/var/lib/kubelet/pods/4dad28d0-454d-434b-b111-137c215aea4a/volumes" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.909340 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" event={"ID":"3dd53027-4438-4022-9ee5-8ed46a8537e3","Type":"ContainerDied","Data":"a561340a77489f4c5d35158b82c3a3c536302bdfb6e9520da586e2c2c89bb552"} Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.909380 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a561340a77489f4c5d35158b82c3a3c536302bdfb6e9520da586e2c2c89bb552" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.909406 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q" Feb 03 09:23:21 crc kubenswrapper[4756]: I0203 09:23:21.913462 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cmfr" event={"ID":"e4384d13-ef52-4709-b694-8ed1628af525","Type":"ContainerStarted","Data":"25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c"} Feb 03 09:23:22 crc kubenswrapper[4756]: I0203 09:23:22.922356 4756 generic.go:334] "Generic (PLEG): container finished" podID="e4384d13-ef52-4709-b694-8ed1628af525" containerID="25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c" exitCode=0 Feb 03 09:23:22 crc kubenswrapper[4756]: I0203 09:23:22.922437 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cmfr" event={"ID":"e4384d13-ef52-4709-b694-8ed1628af525","Type":"ContainerDied","Data":"25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c"} Feb 03 09:23:23 crc kubenswrapper[4756]: I0203 09:23:23.930326 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cmfr" event={"ID":"e4384d13-ef52-4709-b694-8ed1628af525","Type":"ContainerStarted","Data":"95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f"} Feb 03 09:23:23 crc kubenswrapper[4756]: I0203 09:23:23.950365 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2cmfr" podStartSLOduration=2.519470857 podStartE2EDuration="5.950346571s" podCreationTimestamp="2026-02-03 09:23:18 +0000 UTC" firstStartedPulling="2026-02-03 09:23:19.886827229 +0000 UTC m=+791.037294604" lastFinishedPulling="2026-02-03 09:23:23.317702943 +0000 UTC m=+794.468170318" observedRunningTime="2026-02-03 09:23:23.949210775 +0000 UTC m=+795.099678150" watchObservedRunningTime="2026-02-03 09:23:23.950346571 +0000 UTC m=+795.100813956" Feb 03 09:23:28 crc kubenswrapper[4756]: I0203 09:23:28.565291 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:28 crc kubenswrapper[4756]: I0203 09:23:28.565901 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:29 crc kubenswrapper[4756]: I0203 09:23:29.603414 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2cmfr" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="registry-server" probeResult="failure" output=< Feb 03 09:23:29 crc kubenswrapper[4756]: timeout: failed to connect service ":50051" within 1s Feb 03 09:23:29 crc kubenswrapper[4756]: > Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.871587 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k"] Feb 03 09:23:32 crc kubenswrapper[4756]: E0203 09:23:32.871977 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerName="pull" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.871988 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerName="pull" Feb 03 09:23:32 crc kubenswrapper[4756]: E0203 09:23:32.872005 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerName="util" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.872011 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerName="util" Feb 03 09:23:32 crc kubenswrapper[4756]: E0203 09:23:32.872024 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerName="extract" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.872029 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerName="extract" Feb 03 09:23:32 crc kubenswrapper[4756]: E0203 09:23:32.872040 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dad28d0-454d-434b-b111-137c215aea4a" containerName="console" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.872045 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dad28d0-454d-434b-b111-137c215aea4a" containerName="console" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.872138 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dad28d0-454d-434b-b111-137c215aea4a" containerName="console" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.872146 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd53027-4438-4022-9ee5-8ed46a8537e3" containerName="extract" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.872527 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.875912 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.876844 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.877028 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-kwzxd" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.877359 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.877533 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.888430 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k"] Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.919565 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc59b7e7-12f3-4501-94f9-07f3338d7835-webhook-cert\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.919648 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fcbw\" (UniqueName: \"kubernetes.io/projected/fc59b7e7-12f3-4501-94f9-07f3338d7835-kube-api-access-8fcbw\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:32 crc kubenswrapper[4756]: I0203 09:23:32.919714 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fc59b7e7-12f3-4501-94f9-07f3338d7835-apiservice-cert\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.020731 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc59b7e7-12f3-4501-94f9-07f3338d7835-webhook-cert\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.020792 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fcbw\" (UniqueName: \"kubernetes.io/projected/fc59b7e7-12f3-4501-94f9-07f3338d7835-kube-api-access-8fcbw\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.020849 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fc59b7e7-12f3-4501-94f9-07f3338d7835-apiservice-cert\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.026840 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fc59b7e7-12f3-4501-94f9-07f3338d7835-apiservice-cert\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.026880 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc59b7e7-12f3-4501-94f9-07f3338d7835-webhook-cert\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.044148 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fcbw\" (UniqueName: \"kubernetes.io/projected/fc59b7e7-12f3-4501-94f9-07f3338d7835-kube-api-access-8fcbw\") pod \"metallb-operator-controller-manager-576684d48c-mkg4k\" (UID: \"fc59b7e7-12f3-4501-94f9-07f3338d7835\") " pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.125167 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x"] Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.126216 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.129642 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.130908 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.131253 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-q76vm" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.148968 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x"] Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.223273 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12aaddd9-3223-4da5-8040-4149116fede4-apiservice-cert\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.223315 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2b5w\" (UniqueName: \"kubernetes.io/projected/12aaddd9-3223-4da5-8040-4149116fede4-kube-api-access-n2b5w\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.223376 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12aaddd9-3223-4da5-8040-4149116fede4-webhook-cert\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.228478 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.324136 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12aaddd9-3223-4da5-8040-4149116fede4-apiservice-cert\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.324169 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2b5w\" (UniqueName: \"kubernetes.io/projected/12aaddd9-3223-4da5-8040-4149116fede4-kube-api-access-n2b5w\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.324229 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12aaddd9-3223-4da5-8040-4149116fede4-webhook-cert\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.330119 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12aaddd9-3223-4da5-8040-4149116fede4-apiservice-cert\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.335416 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12aaddd9-3223-4da5-8040-4149116fede4-webhook-cert\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.345578 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2b5w\" (UniqueName: \"kubernetes.io/projected/12aaddd9-3223-4da5-8040-4149116fede4-kube-api-access-n2b5w\") pod \"metallb-operator-webhook-server-87484dcc8-cbd6x\" (UID: \"12aaddd9-3223-4da5-8040-4149116fede4\") " pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.443034 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.730423 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k"] Feb 03 09:23:33 crc kubenswrapper[4756]: W0203 09:23:33.739545 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc59b7e7_12f3_4501_94f9_07f3338d7835.slice/crio-d7b3e581ceab525d5f04e7babb152bf206a304e03742db19b00dd3e8d26ef3c9 WatchSource:0}: Error finding container d7b3e581ceab525d5f04e7babb152bf206a304e03742db19b00dd3e8d26ef3c9: Status 404 returned error can't find the container with id d7b3e581ceab525d5f04e7babb152bf206a304e03742db19b00dd3e8d26ef3c9 Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.770277 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x"] Feb 03 09:23:33 crc kubenswrapper[4756]: W0203 09:23:33.776904 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12aaddd9_3223_4da5_8040_4149116fede4.slice/crio-60c0512275e907cd567939f1dd160918d6e4f8b9c9b61dcd189c33715389c091 WatchSource:0}: Error finding container 60c0512275e907cd567939f1dd160918d6e4f8b9c9b61dcd189c33715389c091: Status 404 returned error can't find the container with id 60c0512275e907cd567939f1dd160918d6e4f8b9c9b61dcd189c33715389c091 Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.986225 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" event={"ID":"12aaddd9-3223-4da5-8040-4149116fede4","Type":"ContainerStarted","Data":"60c0512275e907cd567939f1dd160918d6e4f8b9c9b61dcd189c33715389c091"} Feb 03 09:23:33 crc kubenswrapper[4756]: I0203 09:23:33.987434 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" event={"ID":"fc59b7e7-12f3-4501-94f9-07f3338d7835","Type":"ContainerStarted","Data":"d7b3e581ceab525d5f04e7babb152bf206a304e03742db19b00dd3e8d26ef3c9"} Feb 03 09:23:37 crc kubenswrapper[4756]: I0203 09:23:37.007052 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" event={"ID":"fc59b7e7-12f3-4501-94f9-07f3338d7835","Type":"ContainerStarted","Data":"3e71182364846723693b76f34dc9a8269e5f06b41ba9d7620cc166a592867627"} Feb 03 09:23:37 crc kubenswrapper[4756]: I0203 09:23:37.007703 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:23:37 crc kubenswrapper[4756]: I0203 09:23:37.035884 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" podStartSLOduration=2.318243394 podStartE2EDuration="5.03586818s" podCreationTimestamp="2026-02-03 09:23:32 +0000 UTC" firstStartedPulling="2026-02-03 09:23:33.74190926 +0000 UTC m=+804.892376645" lastFinishedPulling="2026-02-03 09:23:36.459534056 +0000 UTC m=+807.610001431" observedRunningTime="2026-02-03 09:23:37.031594988 +0000 UTC m=+808.182062363" watchObservedRunningTime="2026-02-03 09:23:37.03586818 +0000 UTC m=+808.186335555" Feb 03 09:23:38 crc kubenswrapper[4756]: I0203 09:23:38.620692 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:38 crc kubenswrapper[4756]: I0203 09:23:38.670600 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:39 crc kubenswrapper[4756]: I0203 09:23:39.022528 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" event={"ID":"12aaddd9-3223-4da5-8040-4149116fede4","Type":"ContainerStarted","Data":"578305f3bd3a61e9c99bbc378d25db10e80fc1dc3f2e8e98334c8ba9e907be57"} Feb 03 09:23:39 crc kubenswrapper[4756]: I0203 09:23:39.022857 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:23:39 crc kubenswrapper[4756]: I0203 09:23:39.042523 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" podStartSLOduration=1.355955899 podStartE2EDuration="6.042503659s" podCreationTimestamp="2026-02-03 09:23:33 +0000 UTC" firstStartedPulling="2026-02-03 09:23:33.779992496 +0000 UTC m=+804.930459871" lastFinishedPulling="2026-02-03 09:23:38.466540256 +0000 UTC m=+809.617007631" observedRunningTime="2026-02-03 09:23:39.038156124 +0000 UTC m=+810.188623509" watchObservedRunningTime="2026-02-03 09:23:39.042503659 +0000 UTC m=+810.192971054" Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.012267 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2cmfr"] Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.012883 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2cmfr" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="registry-server" containerID="cri-o://95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f" gracePeriod=2 Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.584409 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.643572 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lfnz\" (UniqueName: \"kubernetes.io/projected/e4384d13-ef52-4709-b694-8ed1628af525-kube-api-access-9lfnz\") pod \"e4384d13-ef52-4709-b694-8ed1628af525\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.643654 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-utilities\") pod \"e4384d13-ef52-4709-b694-8ed1628af525\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.643691 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-catalog-content\") pod \"e4384d13-ef52-4709-b694-8ed1628af525\" (UID: \"e4384d13-ef52-4709-b694-8ed1628af525\") " Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.645094 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-utilities" (OuterVolumeSpecName: "utilities") pod "e4384d13-ef52-4709-b694-8ed1628af525" (UID: "e4384d13-ef52-4709-b694-8ed1628af525"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.648969 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4384d13-ef52-4709-b694-8ed1628af525-kube-api-access-9lfnz" (OuterVolumeSpecName: "kube-api-access-9lfnz") pod "e4384d13-ef52-4709-b694-8ed1628af525" (UID: "e4384d13-ef52-4709-b694-8ed1628af525"). InnerVolumeSpecName "kube-api-access-9lfnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.745830 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lfnz\" (UniqueName: \"kubernetes.io/projected/e4384d13-ef52-4709-b694-8ed1628af525-kube-api-access-9lfnz\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.745875 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.762889 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4384d13-ef52-4709-b694-8ed1628af525" (UID: "e4384d13-ef52-4709-b694-8ed1628af525"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:23:41 crc kubenswrapper[4756]: I0203 09:23:41.847248 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4384d13-ef52-4709-b694-8ed1628af525-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.037401 4756 generic.go:334] "Generic (PLEG): container finished" podID="e4384d13-ef52-4709-b694-8ed1628af525" containerID="95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f" exitCode=0 Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.037455 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cmfr" event={"ID":"e4384d13-ef52-4709-b694-8ed1628af525","Type":"ContainerDied","Data":"95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f"} Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.037481 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cmfr" event={"ID":"e4384d13-ef52-4709-b694-8ed1628af525","Type":"ContainerDied","Data":"914c33ded8cd990e409ba346ad699cae014626fd64e69ab4cd7c3bd375854a99"} Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.037496 4756 scope.go:117] "RemoveContainer" containerID="95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.037605 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cmfr" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.069127 4756 scope.go:117] "RemoveContainer" containerID="25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.079360 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2cmfr"] Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.082076 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2cmfr"] Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.092877 4756 scope.go:117] "RemoveContainer" containerID="0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.118331 4756 scope.go:117] "RemoveContainer" containerID="95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f" Feb 03 09:23:42 crc kubenswrapper[4756]: E0203 09:23:42.118871 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f\": container with ID starting with 95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f not found: ID does not exist" containerID="95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.118913 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f"} err="failed to get container status \"95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f\": rpc error: code = NotFound desc = could not find container \"95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f\": container with ID starting with 95552087bd7a0c4fb566f263b38701ff16eb696d971aed75e708473d457df72f not found: ID does not exist" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.118933 4756 scope.go:117] "RemoveContainer" containerID="25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c" Feb 03 09:23:42 crc kubenswrapper[4756]: E0203 09:23:42.119465 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c\": container with ID starting with 25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c not found: ID does not exist" containerID="25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.119506 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c"} err="failed to get container status \"25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c\": rpc error: code = NotFound desc = could not find container \"25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c\": container with ID starting with 25a810c70503baa9408c985cf1e24253f1c39bce80051caec3cbc8cb7e7bae5c not found: ID does not exist" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.119526 4756 scope.go:117] "RemoveContainer" containerID="0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575" Feb 03 09:23:42 crc kubenswrapper[4756]: E0203 09:23:42.120886 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575\": container with ID starting with 0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575 not found: ID does not exist" containerID="0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575" Feb 03 09:23:42 crc kubenswrapper[4756]: I0203 09:23:42.120920 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575"} err="failed to get container status \"0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575\": rpc error: code = NotFound desc = could not find container \"0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575\": container with ID starting with 0b7d11505c50eaf40bdfb38445a8960d1088173a3ed1bb99d70e413a50b2c575 not found: ID does not exist" Feb 03 09:23:43 crc kubenswrapper[4756]: I0203 09:23:43.623419 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4384d13-ef52-4709-b694-8ed1628af525" path="/var/lib/kubelet/pods/e4384d13-ef52-4709-b694-8ed1628af525/volumes" Feb 03 09:23:53 crc kubenswrapper[4756]: I0203 09:23:53.447585 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-87484dcc8-cbd6x" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.231429 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-576684d48c-mkg4k" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.955139 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx"] Feb 03 09:24:13 crc kubenswrapper[4756]: E0203 09:24:13.955762 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="registry-server" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.955864 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="registry-server" Feb 03 09:24:13 crc kubenswrapper[4756]: E0203 09:24:13.955953 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="extract-utilities" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.956027 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="extract-utilities" Feb 03 09:24:13 crc kubenswrapper[4756]: E0203 09:24:13.956096 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="extract-content" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.956162 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="extract-content" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.956355 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4384d13-ef52-4709-b694-8ed1628af525" containerName="registry-server" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.956962 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:13 crc kubenswrapper[4756]: W0203 09:24:13.958453 4756 reflector.go:561] object-"metallb-system"/"frr-k8s-daemon-dockercfg-bvx7j": failed to list *v1.Secret: secrets "frr-k8s-daemon-dockercfg-bvx7j" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Feb 03 09:24:13 crc kubenswrapper[4756]: E0203 09:24:13.958493 4756 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-daemon-dockercfg-bvx7j\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-daemon-dockercfg-bvx7j\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 03 09:24:13 crc kubenswrapper[4756]: W0203 09:24:13.960124 4756 reflector.go:561] object-"metallb-system"/"frr-k8s-webhook-server-cert": failed to list *v1.Secret: secrets "frr-k8s-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Feb 03 09:24:13 crc kubenswrapper[4756]: E0203 09:24:13.960160 4756 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.986522 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-88fxn"] Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.989254 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.992066 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.995754 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 03 09:24:13 crc kubenswrapper[4756]: I0203 09:24:13.996348 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx"] Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.107558 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-reloader\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.107852 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c4e946d8-45d6-4706-982c-32d811489a5a-frr-startup\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.107888 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-frr-conf\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.107906 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x89sq\" (UniqueName: \"kubernetes.io/projected/c4e946d8-45d6-4706-982c-32d811489a5a-kube-api-access-x89sq\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.107921 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-frr-sockets\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.107956 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-bwmnx\" (UID: \"bcaa1bc4-f6a5-4216-8581-13cbf07c2504\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.107975 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tgbm\" (UniqueName: \"kubernetes.io/projected/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-kube-api-access-7tgbm\") pod \"frr-k8s-webhook-server-7df86c4f6c-bwmnx\" (UID: \"bcaa1bc4-f6a5-4216-8581-13cbf07c2504\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.107993 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-metrics\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.108008 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4e946d8-45d6-4706-982c-32d811489a5a-metrics-certs\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.113116 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-44db6"] Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.114263 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.115624 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.116151 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-z5bb4" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.117269 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.117858 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.129939 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-dc6ld"] Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.130765 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.132146 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.154892 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-dc6ld"] Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.208980 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-reloader\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209019 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c4e946d8-45d6-4706-982c-32d811489a5a-frr-startup\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209052 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-frr-sockets\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209066 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-frr-conf\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209093 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x89sq\" (UniqueName: \"kubernetes.io/projected/c4e946d8-45d6-4706-982c-32d811489a5a-kube-api-access-x89sq\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209125 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-bwmnx\" (UID: \"bcaa1bc4-f6a5-4216-8581-13cbf07c2504\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209146 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tgbm\" (UniqueName: \"kubernetes.io/projected/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-kube-api-access-7tgbm\") pod \"frr-k8s-webhook-server-7df86c4f6c-bwmnx\" (UID: \"bcaa1bc4-f6a5-4216-8581-13cbf07c2504\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209165 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-metrics\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209179 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4e946d8-45d6-4706-982c-32d811489a5a-metrics-certs\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: E0203 09:24:14.209290 4756 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 03 09:24:14 crc kubenswrapper[4756]: E0203 09:24:14.209334 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4e946d8-45d6-4706-982c-32d811489a5a-metrics-certs podName:c4e946d8-45d6-4706-982c-32d811489a5a nodeName:}" failed. No retries permitted until 2026-02-03 09:24:14.709318836 +0000 UTC m=+845.859786211 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4e946d8-45d6-4706-982c-32d811489a5a-metrics-certs") pod "frr-k8s-88fxn" (UID: "c4e946d8-45d6-4706-982c-32d811489a5a") : secret "frr-k8s-certs-secret" not found Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.209874 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-reloader\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.210560 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c4e946d8-45d6-4706-982c-32d811489a5a-frr-startup\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.210757 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-frr-sockets\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.210935 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-frr-conf\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.211383 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c4e946d8-45d6-4706-982c-32d811489a5a-metrics\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.229139 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x89sq\" (UniqueName: \"kubernetes.io/projected/c4e946d8-45d6-4706-982c-32d811489a5a-kube-api-access-x89sq\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.232196 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tgbm\" (UniqueName: \"kubernetes.io/projected/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-kube-api-access-7tgbm\") pod \"frr-k8s-webhook-server-7df86c4f6c-bwmnx\" (UID: \"bcaa1bc4-f6a5-4216-8581-13cbf07c2504\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.310690 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16bed8ad-994f-4826-b43d-bc8ef79bce7e-cert\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.310746 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-metrics-certs\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.310796 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.310836 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16bed8ad-994f-4826-b43d-bc8ef79bce7e-metrics-certs\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.310867 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jhr9\" (UniqueName: \"kubernetes.io/projected/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-kube-api-access-5jhr9\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.310898 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr6wb\" (UniqueName: \"kubernetes.io/projected/16bed8ad-994f-4826-b43d-bc8ef79bce7e-kube-api-access-nr6wb\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.310923 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-metallb-excludel2\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.412214 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-metrics-certs\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.412336 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: E0203 09:24:14.412500 4756 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 03 09:24:14 crc kubenswrapper[4756]: E0203 09:24:14.412563 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist podName:98d5a2ec-16f8-4654-b5ba-0c32b08d0fda nodeName:}" failed. No retries permitted until 2026-02-03 09:24:14.912543205 +0000 UTC m=+846.063010580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist") pod "speaker-44db6" (UID: "98d5a2ec-16f8-4654-b5ba-0c32b08d0fda") : secret "metallb-memberlist" not found Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.412374 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16bed8ad-994f-4826-b43d-bc8ef79bce7e-metrics-certs\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.413476 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jhr9\" (UniqueName: \"kubernetes.io/projected/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-kube-api-access-5jhr9\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.413535 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr6wb\" (UniqueName: \"kubernetes.io/projected/16bed8ad-994f-4826-b43d-bc8ef79bce7e-kube-api-access-nr6wb\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.413558 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-metallb-excludel2\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.413590 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16bed8ad-994f-4826-b43d-bc8ef79bce7e-cert\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.414368 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-metallb-excludel2\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.416646 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.416859 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-metrics-certs\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.416906 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16bed8ad-994f-4826-b43d-bc8ef79bce7e-metrics-certs\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.427155 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16bed8ad-994f-4826-b43d-bc8ef79bce7e-cert\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.445630 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr6wb\" (UniqueName: \"kubernetes.io/projected/16bed8ad-994f-4826-b43d-bc8ef79bce7e-kube-api-access-nr6wb\") pod \"controller-6968d8fdc4-dc6ld\" (UID: \"16bed8ad-994f-4826-b43d-bc8ef79bce7e\") " pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.448840 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jhr9\" (UniqueName: \"kubernetes.io/projected/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-kube-api-access-5jhr9\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.716544 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4e946d8-45d6-4706-982c-32d811489a5a-metrics-certs\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.719700 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4e946d8-45d6-4706-982c-32d811489a5a-metrics-certs\") pod \"frr-k8s-88fxn\" (UID: \"c4e946d8-45d6-4706-982c-32d811489a5a\") " pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.744088 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.768361 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-bvx7j" Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.919220 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:14 crc kubenswrapper[4756]: E0203 09:24:14.919624 4756 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 03 09:24:14 crc kubenswrapper[4756]: E0203 09:24:14.919670 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist podName:98d5a2ec-16f8-4654-b5ba-0c32b08d0fda nodeName:}" failed. No retries permitted until 2026-02-03 09:24:15.919656038 +0000 UTC m=+847.070123413 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist") pod "speaker-44db6" (UID: "98d5a2ec-16f8-4654-b5ba-0c32b08d0fda") : secret "metallb-memberlist" not found Feb 03 09:24:14 crc kubenswrapper[4756]: I0203 09:24:14.938731 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.152806 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-dc6ld"] Feb 03 09:24:15 crc kubenswrapper[4756]: W0203 09:24:15.158740 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16bed8ad_994f_4826_b43d_bc8ef79bce7e.slice/crio-1157d3c75ca00c8239e5d63e18b7020a11f436b300884493899f2d44afcdf886 WatchSource:0}: Error finding container 1157d3c75ca00c8239e5d63e18b7020a11f436b300884493899f2d44afcdf886: Status 404 returned error can't find the container with id 1157d3c75ca00c8239e5d63e18b7020a11f436b300884493899f2d44afcdf886 Feb 03 09:24:15 crc kubenswrapper[4756]: E0203 09:24:15.211203 4756 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: failed to sync secret cache: timed out waiting for the condition Feb 03 09:24:15 crc kubenswrapper[4756]: E0203 09:24:15.211272 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-cert podName:bcaa1bc4-f6a5-4216-8581-13cbf07c2504 nodeName:}" failed. No retries permitted until 2026-02-03 09:24:15.711254438 +0000 UTC m=+846.861721813 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-cert") pod "frr-k8s-webhook-server-7df86c4f6c-bwmnx" (UID: "bcaa1bc4-f6a5-4216-8581-13cbf07c2504") : failed to sync secret cache: timed out waiting for the condition Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.211941 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-dc6ld" event={"ID":"16bed8ad-994f-4826-b43d-bc8ef79bce7e","Type":"ContainerStarted","Data":"1157d3c75ca00c8239e5d63e18b7020a11f436b300884493899f2d44afcdf886"} Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.212948 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerStarted","Data":"720ef81a37c271f47c896a7889612fc1d5243f0e2664088c40660dac8f31df20"} Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.277419 4756 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.731312 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-bwmnx\" (UID: \"bcaa1bc4-f6a5-4216-8581-13cbf07c2504\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.740233 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bcaa1bc4-f6a5-4216-8581-13cbf07c2504-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-bwmnx\" (UID: \"bcaa1bc4-f6a5-4216-8581-13cbf07c2504\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.770039 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.934716 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:15 crc kubenswrapper[4756]: I0203 09:24:15.939471 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/98d5a2ec-16f8-4654-b5ba-0c32b08d0fda-memberlist\") pod \"speaker-44db6\" (UID: \"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda\") " pod="metallb-system/speaker-44db6" Feb 03 09:24:16 crc kubenswrapper[4756]: I0203 09:24:16.178535 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx"] Feb 03 09:24:16 crc kubenswrapper[4756]: I0203 09:24:16.229278 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-dc6ld" event={"ID":"16bed8ad-994f-4826-b43d-bc8ef79bce7e","Type":"ContainerStarted","Data":"03c76b04cc04731f6f83c404d299f66681522f4496925be7a5b99fb786e61924"} Feb 03 09:24:16 crc kubenswrapper[4756]: I0203 09:24:16.229326 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-dc6ld" event={"ID":"16bed8ad-994f-4826-b43d-bc8ef79bce7e","Type":"ContainerStarted","Data":"b966dfc2f2a1cfdd78e97c956e3e0d7052da06c58cdd3b5e5b4b7768ffee8b67"} Feb 03 09:24:16 crc kubenswrapper[4756]: I0203 09:24:16.230206 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:16 crc kubenswrapper[4756]: I0203 09:24:16.231426 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" event={"ID":"bcaa1bc4-f6a5-4216-8581-13cbf07c2504","Type":"ContainerStarted","Data":"ad769de003c4f7db1d64c1a8cdf66c0b28bd5ff3f51d940383a1e4270cd9db11"} Feb 03 09:24:16 crc kubenswrapper[4756]: I0203 09:24:16.231592 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-44db6" Feb 03 09:24:16 crc kubenswrapper[4756]: I0203 09:24:16.260390 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-dc6ld" podStartSLOduration=2.2603736850000002 podStartE2EDuration="2.260373685s" podCreationTimestamp="2026-02-03 09:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:24:16.259807107 +0000 UTC m=+847.410274502" watchObservedRunningTime="2026-02-03 09:24:16.260373685 +0000 UTC m=+847.410841050" Feb 03 09:24:17 crc kubenswrapper[4756]: I0203 09:24:17.241028 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-44db6" event={"ID":"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda","Type":"ContainerStarted","Data":"01b1febbf7e0b877fcec7203d3b25ce618a33ce6a7c9c5ed5f6c8c893e726f73"} Feb 03 09:24:17 crc kubenswrapper[4756]: I0203 09:24:17.241345 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-44db6" event={"ID":"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda","Type":"ContainerStarted","Data":"341f00344fcae03efef2f9362ff699b4d764a50118803ed342ab61c637a4358c"} Feb 03 09:24:17 crc kubenswrapper[4756]: I0203 09:24:17.241366 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-44db6" event={"ID":"98d5a2ec-16f8-4654-b5ba-0c32b08d0fda","Type":"ContainerStarted","Data":"3abe3c5484d91ad5ce85ec0c3df8d2b0a1d205f56f5622825ada4867469116b7"} Feb 03 09:24:17 crc kubenswrapper[4756]: I0203 09:24:17.241548 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-44db6" Feb 03 09:24:17 crc kubenswrapper[4756]: I0203 09:24:17.280267 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-44db6" podStartSLOduration=3.280242067 podStartE2EDuration="3.280242067s" podCreationTimestamp="2026-02-03 09:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:24:17.262869474 +0000 UTC m=+848.413336869" watchObservedRunningTime="2026-02-03 09:24:17.280242067 +0000 UTC m=+848.430709442" Feb 03 09:24:22 crc kubenswrapper[4756]: I0203 09:24:22.276563 4756 generic.go:334] "Generic (PLEG): container finished" podID="c4e946d8-45d6-4706-982c-32d811489a5a" containerID="adc106e86a0fad3c521aa55753332005a239e0780cc943467bf97c62aaa2f197" exitCode=0 Feb 03 09:24:22 crc kubenswrapper[4756]: I0203 09:24:22.276611 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerDied","Data":"adc106e86a0fad3c521aa55753332005a239e0780cc943467bf97c62aaa2f197"} Feb 03 09:24:22 crc kubenswrapper[4756]: I0203 09:24:22.277853 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" event={"ID":"bcaa1bc4-f6a5-4216-8581-13cbf07c2504","Type":"ContainerStarted","Data":"4a80a32357c221959b7a500732ef60d2e6dae9d4dce511a241a51d527e8a46ea"} Feb 03 09:24:22 crc kubenswrapper[4756]: I0203 09:24:22.278171 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:22 crc kubenswrapper[4756]: I0203 09:24:22.327071 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" podStartSLOduration=3.522736982 podStartE2EDuration="9.327051609s" podCreationTimestamp="2026-02-03 09:24:13 +0000 UTC" firstStartedPulling="2026-02-03 09:24:16.209629889 +0000 UTC m=+847.360097264" lastFinishedPulling="2026-02-03 09:24:22.013944526 +0000 UTC m=+853.164411891" observedRunningTime="2026-02-03 09:24:22.314038212 +0000 UTC m=+853.464505597" watchObservedRunningTime="2026-02-03 09:24:22.327051609 +0000 UTC m=+853.477518984" Feb 03 09:24:23 crc kubenswrapper[4756]: I0203 09:24:23.284423 4756 generic.go:334] "Generic (PLEG): container finished" podID="c4e946d8-45d6-4706-982c-32d811489a5a" containerID="520f581c934e0c009f04d085aa696cc54275fea7b95dcc12511c743cef7cf8dc" exitCode=0 Feb 03 09:24:23 crc kubenswrapper[4756]: I0203 09:24:23.285765 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerDied","Data":"520f581c934e0c009f04d085aa696cc54275fea7b95dcc12511c743cef7cf8dc"} Feb 03 09:24:24 crc kubenswrapper[4756]: I0203 09:24:24.293825 4756 generic.go:334] "Generic (PLEG): container finished" podID="c4e946d8-45d6-4706-982c-32d811489a5a" containerID="4bf82211ef5b73f63ceef1c3a9f01123ab5e78b055507c77cb80d1ecaa033519" exitCode=0 Feb 03 09:24:24 crc kubenswrapper[4756]: I0203 09:24:24.293864 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerDied","Data":"4bf82211ef5b73f63ceef1c3a9f01123ab5e78b055507c77cb80d1ecaa033519"} Feb 03 09:24:25 crc kubenswrapper[4756]: I0203 09:24:25.306172 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerStarted","Data":"266f46999463fe4128d2e65fdac2242bf02a1b4366d9184e6070de60a6e5b5f9"} Feb 03 09:24:25 crc kubenswrapper[4756]: I0203 09:24:25.306397 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerStarted","Data":"ea6aada26bedb5a1b89e8b4349dc0099504e30dfb77b0c20ca5265c7042fccae"} Feb 03 09:24:25 crc kubenswrapper[4756]: I0203 09:24:25.306407 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerStarted","Data":"0fd69754407bfcfa3e5b784b6402580da480ea544d61988c49ebc52777b69ecf"} Feb 03 09:24:25 crc kubenswrapper[4756]: I0203 09:24:25.306417 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerStarted","Data":"fb3203d91b9970c5c45800bd9ef775fba5347691bd80b3f51ebcf050b944ac73"} Feb 03 09:24:25 crc kubenswrapper[4756]: I0203 09:24:25.306425 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerStarted","Data":"27e1178843fec7da8579b00110fc2da2c30d5e7882753ea2ef082d45a1fa632c"} Feb 03 09:24:25 crc kubenswrapper[4756]: I0203 09:24:25.306433 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-88fxn" event={"ID":"c4e946d8-45d6-4706-982c-32d811489a5a","Type":"ContainerStarted","Data":"86e3594e20f018032226065e05cca343271ce8a9f4335e9adf8023c8a19a2066"} Feb 03 09:24:25 crc kubenswrapper[4756]: I0203 09:24:25.306461 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:25 crc kubenswrapper[4756]: I0203 09:24:25.330794 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-88fxn" podStartSLOduration=5.382485193 podStartE2EDuration="12.330781241s" podCreationTimestamp="2026-02-03 09:24:13 +0000 UTC" firstStartedPulling="2026-02-03 09:24:15.046398157 +0000 UTC m=+846.196865532" lastFinishedPulling="2026-02-03 09:24:21.994694195 +0000 UTC m=+853.145161580" observedRunningTime="2026-02-03 09:24:25.327579531 +0000 UTC m=+856.478046906" watchObservedRunningTime="2026-02-03 09:24:25.330781241 +0000 UTC m=+856.481248616" Feb 03 09:24:26 crc kubenswrapper[4756]: I0203 09:24:26.238866 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-44db6" Feb 03 09:24:28 crc kubenswrapper[4756]: I0203 09:24:28.986454 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-z8gwt"] Feb 03 09:24:28 crc kubenswrapper[4756]: I0203 09:24:28.987878 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z8gwt" Feb 03 09:24:28 crc kubenswrapper[4756]: I0203 09:24:28.993787 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 03 09:24:28 crc kubenswrapper[4756]: I0203 09:24:28.994092 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 03 09:24:28 crc kubenswrapper[4756]: I0203 09:24:28.994341 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-lq2xq" Feb 03 09:24:29 crc kubenswrapper[4756]: I0203 09:24:29.013536 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x499j\" (UniqueName: \"kubernetes.io/projected/17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6-kube-api-access-x499j\") pod \"openstack-operator-index-z8gwt\" (UID: \"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6\") " pod="openstack-operators/openstack-operator-index-z8gwt" Feb 03 09:24:29 crc kubenswrapper[4756]: I0203 09:24:29.018365 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z8gwt"] Feb 03 09:24:29 crc kubenswrapper[4756]: I0203 09:24:29.114810 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x499j\" (UniqueName: \"kubernetes.io/projected/17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6-kube-api-access-x499j\") pod \"openstack-operator-index-z8gwt\" (UID: \"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6\") " pod="openstack-operators/openstack-operator-index-z8gwt" Feb 03 09:24:29 crc kubenswrapper[4756]: I0203 09:24:29.136156 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x499j\" (UniqueName: \"kubernetes.io/projected/17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6-kube-api-access-x499j\") pod \"openstack-operator-index-z8gwt\" (UID: \"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6\") " pod="openstack-operators/openstack-operator-index-z8gwt" Feb 03 09:24:29 crc kubenswrapper[4756]: I0203 09:24:29.331267 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z8gwt" Feb 03 09:24:29 crc kubenswrapper[4756]: I0203 09:24:29.732628 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z8gwt"] Feb 03 09:24:29 crc kubenswrapper[4756]: I0203 09:24:29.939663 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:30 crc kubenswrapper[4756]: I0203 09:24:30.000022 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:30 crc kubenswrapper[4756]: I0203 09:24:30.349383 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z8gwt" event={"ID":"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6","Type":"ContainerStarted","Data":"f29b64caa46982342a5c6394c3a5b036072770c5fd8ca825e8a5cfab318cab24"} Feb 03 09:24:32 crc kubenswrapper[4756]: I0203 09:24:32.358165 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-z8gwt"] Feb 03 09:24:32 crc kubenswrapper[4756]: I0203 09:24:32.361933 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z8gwt" event={"ID":"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6","Type":"ContainerStarted","Data":"052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009"} Feb 03 09:24:32 crc kubenswrapper[4756]: I0203 09:24:32.381614 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-z8gwt" podStartSLOduration=2.460016897 podStartE2EDuration="4.381593691s" podCreationTimestamp="2026-02-03 09:24:28 +0000 UTC" firstStartedPulling="2026-02-03 09:24:29.743484751 +0000 UTC m=+860.893952126" lastFinishedPulling="2026-02-03 09:24:31.665061545 +0000 UTC m=+862.815528920" observedRunningTime="2026-02-03 09:24:32.380053384 +0000 UTC m=+863.530520779" watchObservedRunningTime="2026-02-03 09:24:32.381593691 +0000 UTC m=+863.532061076" Feb 03 09:24:32 crc kubenswrapper[4756]: I0203 09:24:32.963265 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-r84l6"] Feb 03 09:24:32 crc kubenswrapper[4756]: I0203 09:24:32.964959 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:32 crc kubenswrapper[4756]: I0203 09:24:32.968000 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r84l6"] Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.063729 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p52k\" (UniqueName: \"kubernetes.io/projected/9dc52ada-64d7-45a9-b880-ded26c6ec2f6-kube-api-access-7p52k\") pod \"openstack-operator-index-r84l6\" (UID: \"9dc52ada-64d7-45a9-b880-ded26c6ec2f6\") " pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.164843 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p52k\" (UniqueName: \"kubernetes.io/projected/9dc52ada-64d7-45a9-b880-ded26c6ec2f6-kube-api-access-7p52k\") pod \"openstack-operator-index-r84l6\" (UID: \"9dc52ada-64d7-45a9-b880-ded26c6ec2f6\") " pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.194537 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p52k\" (UniqueName: \"kubernetes.io/projected/9dc52ada-64d7-45a9-b880-ded26c6ec2f6-kube-api-access-7p52k\") pod \"openstack-operator-index-r84l6\" (UID: \"9dc52ada-64d7-45a9-b880-ded26c6ec2f6\") " pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.286100 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.367354 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-z8gwt" podUID="17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6" containerName="registry-server" containerID="cri-o://052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009" gracePeriod=2 Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.687907 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z8gwt" Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.689193 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r84l6"] Feb 03 09:24:33 crc kubenswrapper[4756]: W0203 09:24:33.695002 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dc52ada_64d7_45a9_b880_ded26c6ec2f6.slice/crio-92bc9d3314810b01ef5b0b87c213aa1d54f943ec29376dd8080625f6acb27d4d WatchSource:0}: Error finding container 92bc9d3314810b01ef5b0b87c213aa1d54f943ec29376dd8080625f6acb27d4d: Status 404 returned error can't find the container with id 92bc9d3314810b01ef5b0b87c213aa1d54f943ec29376dd8080625f6acb27d4d Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.787677 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x499j\" (UniqueName: \"kubernetes.io/projected/17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6-kube-api-access-x499j\") pod \"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6\" (UID: \"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6\") " Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.792411 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6-kube-api-access-x499j" (OuterVolumeSpecName: "kube-api-access-x499j") pod "17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6" (UID: "17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6"). InnerVolumeSpecName "kube-api-access-x499j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:24:33 crc kubenswrapper[4756]: I0203 09:24:33.889012 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x499j\" (UniqueName: \"kubernetes.io/projected/17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6-kube-api-access-x499j\") on node \"crc\" DevicePath \"\"" Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.374604 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r84l6" event={"ID":"9dc52ada-64d7-45a9-b880-ded26c6ec2f6","Type":"ContainerStarted","Data":"3541cb185d7af09317b8796627b05a5b2a803eddda4813a815fad5e791a73990"} Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.374933 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r84l6" event={"ID":"9dc52ada-64d7-45a9-b880-ded26c6ec2f6","Type":"ContainerStarted","Data":"92bc9d3314810b01ef5b0b87c213aa1d54f943ec29376dd8080625f6acb27d4d"} Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.376286 4756 generic.go:334] "Generic (PLEG): container finished" podID="17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6" containerID="052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009" exitCode=0 Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.376325 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z8gwt" event={"ID":"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6","Type":"ContainerDied","Data":"052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009"} Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.376347 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z8gwt" Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.376360 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z8gwt" event={"ID":"17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6","Type":"ContainerDied","Data":"f29b64caa46982342a5c6394c3a5b036072770c5fd8ca825e8a5cfab318cab24"} Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.376385 4756 scope.go:117] "RemoveContainer" containerID="052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009" Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.395269 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-r84l6" podStartSLOduration=2.34590935 podStartE2EDuration="2.395245502s" podCreationTimestamp="2026-02-03 09:24:32 +0000 UTC" firstStartedPulling="2026-02-03 09:24:33.699928988 +0000 UTC m=+864.850396363" lastFinishedPulling="2026-02-03 09:24:33.74926515 +0000 UTC m=+864.899732515" observedRunningTime="2026-02-03 09:24:34.389598565 +0000 UTC m=+865.540065970" watchObservedRunningTime="2026-02-03 09:24:34.395245502 +0000 UTC m=+865.545712877" Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.401777 4756 scope.go:117] "RemoveContainer" containerID="052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009" Feb 03 09:24:34 crc kubenswrapper[4756]: E0203 09:24:34.402765 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009\": container with ID starting with 052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009 not found: ID does not exist" containerID="052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009" Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.402835 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009"} err="failed to get container status \"052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009\": rpc error: code = NotFound desc = could not find container \"052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009\": container with ID starting with 052ca8bebb9b14f005f35451b2525054cd0786c4675a54ab4d15b999a971d009 not found: ID does not exist" Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.416602 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-z8gwt"] Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.419895 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-z8gwt"] Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.749312 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-dc6ld" Feb 03 09:24:34 crc kubenswrapper[4756]: I0203 09:24:34.942015 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-88fxn" Feb 03 09:24:35 crc kubenswrapper[4756]: I0203 09:24:35.624466 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6" path="/var/lib/kubelet/pods/17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6/volumes" Feb 03 09:24:35 crc kubenswrapper[4756]: I0203 09:24:35.777018 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-bwmnx" Feb 03 09:24:43 crc kubenswrapper[4756]: I0203 09:24:43.286643 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:43 crc kubenswrapper[4756]: I0203 09:24:43.287204 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:43 crc kubenswrapper[4756]: I0203 09:24:43.314303 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:43 crc kubenswrapper[4756]: I0203 09:24:43.458176 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-r84l6" Feb 03 09:24:43 crc kubenswrapper[4756]: I0203 09:24:43.566834 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:24:43 crc kubenswrapper[4756]: I0203 09:24:43.566942 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.017416 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8"] Feb 03 09:24:50 crc kubenswrapper[4756]: E0203 09:24:50.018022 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6" containerName="registry-server" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.018034 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6" containerName="registry-server" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.018136 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="17b8cd12-fd66-4d6e-b3af-c2b3ec7c52e6" containerName="registry-server" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.018925 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.023565 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwpks" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.032052 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8"] Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.129763 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-bundle\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.129823 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-util\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.129860 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg9wx\" (UniqueName: \"kubernetes.io/projected/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-kube-api-access-xg9wx\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.231090 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-bundle\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.231566 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-util\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.231524 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-bundle\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.231643 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg9wx\" (UniqueName: \"kubernetes.io/projected/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-kube-api-access-xg9wx\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.231875 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-util\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.251857 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg9wx\" (UniqueName: \"kubernetes.io/projected/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-kube-api-access-xg9wx\") pod \"198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.387804 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:50 crc kubenswrapper[4756]: I0203 09:24:50.566661 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8"] Feb 03 09:24:51 crc kubenswrapper[4756]: I0203 09:24:51.477888 4756 generic.go:334] "Generic (PLEG): container finished" podID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerID="348556bca1d5ab0b29b57fc5894991e037eb19a1ed9d3e4f0707f5d1263ce376" exitCode=0 Feb 03 09:24:51 crc kubenswrapper[4756]: I0203 09:24:51.477941 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" event={"ID":"1e21c47a-ab4b-4663-ba66-21f3b8c4975b","Type":"ContainerDied","Data":"348556bca1d5ab0b29b57fc5894991e037eb19a1ed9d3e4f0707f5d1263ce376"} Feb 03 09:24:51 crc kubenswrapper[4756]: I0203 09:24:51.477969 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" event={"ID":"1e21c47a-ab4b-4663-ba66-21f3b8c4975b","Type":"ContainerStarted","Data":"d439e2c3d4c0dbc47327cbcc8ab67e74266a29eaebb243ea6d4b4ef51a8c77f4"} Feb 03 09:24:52 crc kubenswrapper[4756]: I0203 09:24:52.496824 4756 generic.go:334] "Generic (PLEG): container finished" podID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerID="06235f5f886e6cd8b1a0ce86ca2fed4112f91ad33a7e6ad35fc897fc1f605601" exitCode=0 Feb 03 09:24:52 crc kubenswrapper[4756]: I0203 09:24:52.497783 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" event={"ID":"1e21c47a-ab4b-4663-ba66-21f3b8c4975b","Type":"ContainerDied","Data":"06235f5f886e6cd8b1a0ce86ca2fed4112f91ad33a7e6ad35fc897fc1f605601"} Feb 03 09:24:54 crc kubenswrapper[4756]: I0203 09:24:54.512157 4756 generic.go:334] "Generic (PLEG): container finished" podID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerID="fcea38beab58f0e373e4229e165f03b7618ecb7f5c9296fc3fbb42b8891dfa80" exitCode=0 Feb 03 09:24:54 crc kubenswrapper[4756]: I0203 09:24:54.512200 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" event={"ID":"1e21c47a-ab4b-4663-ba66-21f3b8c4975b","Type":"ContainerDied","Data":"fcea38beab58f0e373e4229e165f03b7618ecb7f5c9296fc3fbb42b8891dfa80"} Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.765648 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.829085 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-bundle\") pod \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.829159 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg9wx\" (UniqueName: \"kubernetes.io/projected/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-kube-api-access-xg9wx\") pod \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.829220 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-util\") pod \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\" (UID: \"1e21c47a-ab4b-4663-ba66-21f3b8c4975b\") " Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.829763 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-bundle" (OuterVolumeSpecName: "bundle") pod "1e21c47a-ab4b-4663-ba66-21f3b8c4975b" (UID: "1e21c47a-ab4b-4663-ba66-21f3b8c4975b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.836751 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-kube-api-access-xg9wx" (OuterVolumeSpecName: "kube-api-access-xg9wx") pod "1e21c47a-ab4b-4663-ba66-21f3b8c4975b" (UID: "1e21c47a-ab4b-4663-ba66-21f3b8c4975b"). InnerVolumeSpecName "kube-api-access-xg9wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.853344 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-util" (OuterVolumeSpecName: "util") pod "1e21c47a-ab4b-4663-ba66-21f3b8c4975b" (UID: "1e21c47a-ab4b-4663-ba66-21f3b8c4975b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.931106 4756 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-util\") on node \"crc\" DevicePath \"\"" Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.931175 4756 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:24:55 crc kubenswrapper[4756]: I0203 09:24:55.931190 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg9wx\" (UniqueName: \"kubernetes.io/projected/1e21c47a-ab4b-4663-ba66-21f3b8c4975b-kube-api-access-xg9wx\") on node \"crc\" DevicePath \"\"" Feb 03 09:24:56 crc kubenswrapper[4756]: I0203 09:24:56.527711 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" event={"ID":"1e21c47a-ab4b-4663-ba66-21f3b8c4975b","Type":"ContainerDied","Data":"d439e2c3d4c0dbc47327cbcc8ab67e74266a29eaebb243ea6d4b4ef51a8c77f4"} Feb 03 09:24:56 crc kubenswrapper[4756]: I0203 09:24:56.527752 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d439e2c3d4c0dbc47327cbcc8ab67e74266a29eaebb243ea6d4b4ef51a8c77f4" Feb 03 09:24:56 crc kubenswrapper[4756]: I0203 09:24:56.527831 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.184279 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg"] Feb 03 09:24:58 crc kubenswrapper[4756]: E0203 09:24:58.185939 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerName="util" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.186029 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerName="util" Feb 03 09:24:58 crc kubenswrapper[4756]: E0203 09:24:58.186123 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerName="pull" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.186189 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerName="pull" Feb 03 09:24:58 crc kubenswrapper[4756]: E0203 09:24:58.186272 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerName="extract" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.186344 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerName="extract" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.186556 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e21c47a-ab4b-4663-ba66-21f3b8c4975b" containerName="extract" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.187107 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.191687 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-9tdnb" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.242503 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg"] Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.262635 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xcrc\" (UniqueName: \"kubernetes.io/projected/2e4a4670-318d-4489-9a71-a084e2445361-kube-api-access-6xcrc\") pod \"openstack-operator-controller-init-7f67889c79-dwpgg\" (UID: \"2e4a4670-318d-4489-9a71-a084e2445361\") " pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.363575 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xcrc\" (UniqueName: \"kubernetes.io/projected/2e4a4670-318d-4489-9a71-a084e2445361-kube-api-access-6xcrc\") pod \"openstack-operator-controller-init-7f67889c79-dwpgg\" (UID: \"2e4a4670-318d-4489-9a71-a084e2445361\") " pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.382358 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xcrc\" (UniqueName: \"kubernetes.io/projected/2e4a4670-318d-4489-9a71-a084e2445361-kube-api-access-6xcrc\") pod \"openstack-operator-controller-init-7f67889c79-dwpgg\" (UID: \"2e4a4670-318d-4489-9a71-a084e2445361\") " pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.505783 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" Feb 03 09:24:58 crc kubenswrapper[4756]: I0203 09:24:58.759628 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg"] Feb 03 09:24:59 crc kubenswrapper[4756]: I0203 09:24:59.582207 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" event={"ID":"2e4a4670-318d-4489-9a71-a084e2445361","Type":"ContainerStarted","Data":"ae0bc05576463845fb96fd08c571ccbb7ce310f963e37f3971150e597fd1ecfc"} Feb 03 09:25:02 crc kubenswrapper[4756]: I0203 09:25:02.599012 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" event={"ID":"2e4a4670-318d-4489-9a71-a084e2445361","Type":"ContainerStarted","Data":"e5a16ed630cae2561764962c9a51574d99e9daa5b3f5bf7ede77d7b35b271155"} Feb 03 09:25:02 crc kubenswrapper[4756]: I0203 09:25:02.599563 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" Feb 03 09:25:02 crc kubenswrapper[4756]: I0203 09:25:02.625051 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" podStartSLOduration=1.200259118 podStartE2EDuration="4.625035345s" podCreationTimestamp="2026-02-03 09:24:58 +0000 UTC" firstStartedPulling="2026-02-03 09:24:58.766579509 +0000 UTC m=+889.917046884" lastFinishedPulling="2026-02-03 09:25:02.191355746 +0000 UTC m=+893.341823111" observedRunningTime="2026-02-03 09:25:02.620458122 +0000 UTC m=+893.770925507" watchObservedRunningTime="2026-02-03 09:25:02.625035345 +0000 UTC m=+893.775502720" Feb 03 09:25:08 crc kubenswrapper[4756]: I0203 09:25:08.509690 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-7f67889c79-dwpgg" Feb 03 09:25:13 crc kubenswrapper[4756]: I0203 09:25:13.566427 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:25:13 crc kubenswrapper[4756]: I0203 09:25:13.567016 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.602617 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ftt9v"] Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.604345 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.624078 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ftt9v"] Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.640919 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95gw8\" (UniqueName: \"kubernetes.io/projected/64eb5569-95ab-4135-9d4f-4bda9904b174-kube-api-access-95gw8\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.640988 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-utilities\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.641026 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-catalog-content\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.742389 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-catalog-content\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.742508 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95gw8\" (UniqueName: \"kubernetes.io/projected/64eb5569-95ab-4135-9d4f-4bda9904b174-kube-api-access-95gw8\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.742541 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-utilities\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.742974 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-catalog-content\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.743025 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-utilities\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.761903 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95gw8\" (UniqueName: \"kubernetes.io/projected/64eb5569-95ab-4135-9d4f-4bda9904b174-kube-api-access-95gw8\") pod \"certified-operators-ftt9v\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:17 crc kubenswrapper[4756]: I0203 09:25:17.924544 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:18 crc kubenswrapper[4756]: I0203 09:25:18.402334 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ftt9v"] Feb 03 09:25:18 crc kubenswrapper[4756]: I0203 09:25:18.682880 4756 generic.go:334] "Generic (PLEG): container finished" podID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerID="9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e" exitCode=0 Feb 03 09:25:18 crc kubenswrapper[4756]: I0203 09:25:18.682927 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ftt9v" event={"ID":"64eb5569-95ab-4135-9d4f-4bda9904b174","Type":"ContainerDied","Data":"9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e"} Feb 03 09:25:18 crc kubenswrapper[4756]: I0203 09:25:18.682956 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ftt9v" event={"ID":"64eb5569-95ab-4135-9d4f-4bda9904b174","Type":"ContainerStarted","Data":"9f1b30c44f4eef43d1cf1c1e4b3b2bdf3a6a353c607c790c317a48a8df600971"} Feb 03 09:25:19 crc kubenswrapper[4756]: I0203 09:25:19.690166 4756 generic.go:334] "Generic (PLEG): container finished" podID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerID="36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2" exitCode=0 Feb 03 09:25:19 crc kubenswrapper[4756]: I0203 09:25:19.690227 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ftt9v" event={"ID":"64eb5569-95ab-4135-9d4f-4bda9904b174","Type":"ContainerDied","Data":"36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2"} Feb 03 09:25:20 crc kubenswrapper[4756]: I0203 09:25:20.700759 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ftt9v" event={"ID":"64eb5569-95ab-4135-9d4f-4bda9904b174","Type":"ContainerStarted","Data":"8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300"} Feb 03 09:25:27 crc kubenswrapper[4756]: I0203 09:25:27.925432 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:27 crc kubenswrapper[4756]: I0203 09:25:27.925949 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:27 crc kubenswrapper[4756]: I0203 09:25:27.971516 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:27 crc kubenswrapper[4756]: I0203 09:25:27.992229 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ftt9v" podStartSLOduration=9.569602159 podStartE2EDuration="10.992210964s" podCreationTimestamp="2026-02-03 09:25:17 +0000 UTC" firstStartedPulling="2026-02-03 09:25:18.684124852 +0000 UTC m=+909.834592227" lastFinishedPulling="2026-02-03 09:25:20.106733647 +0000 UTC m=+911.257201032" observedRunningTime="2026-02-03 09:25:20.729748431 +0000 UTC m=+911.880215816" watchObservedRunningTime="2026-02-03 09:25:27.992210964 +0000 UTC m=+919.142678339" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.211111 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.212040 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.214104 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-n6sz9" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.216267 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.217116 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.219002 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-f48l4" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.225371 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.226386 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.228972 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dnx5x" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.233404 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.239083 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.247296 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.248273 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.252926 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-wkgj4" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.270506 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.282215 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.289661 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f9xl\" (UniqueName: \"kubernetes.io/projected/6dd1801b-ac57-4d28-9554-ef8ad7148c08-kube-api-access-7f9xl\") pod \"glance-operator-controller-manager-8886f4c47-pgdzr\" (UID: \"6dd1801b-ac57-4d28-9554-ef8ad7148c08\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.289734 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65kzb\" (UniqueName: \"kubernetes.io/projected/a6153ce6-006d-4351-8a9b-aa0abe1cbdad-kube-api-access-65kzb\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-6pbkt\" (UID: \"a6153ce6-006d-4351-8a9b-aa0abe1cbdad\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.289797 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4wrv\" (UniqueName: \"kubernetes.io/projected/cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce-kube-api-access-w4wrv\") pod \"cinder-operator-controller-manager-8d874c8fc-dplk2\" (UID: \"cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.289840 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8699\" (UniqueName: \"kubernetes.io/projected/d74f0516-3682-4b80-beb4-f7237c4daff8-kube-api-access-p8699\") pod \"designate-operator-controller-manager-6d9697b7f4-qnklz\" (UID: \"d74f0516-3682-4b80-beb4-f7237c4daff8\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.297408 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-564mj"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.298266 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.301746 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5g5zx" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.324602 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-564mj"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.332658 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.333451 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.339818 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-crf8q" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.348854 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.349718 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.355221 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.357761 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.361577 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rbw8h" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.367447 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.368380 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.371051 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tpgzb" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.375010 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.388064 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.390547 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4wrv\" (UniqueName: \"kubernetes.io/projected/cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce-kube-api-access-w4wrv\") pod \"cinder-operator-controller-manager-8d874c8fc-dplk2\" (UID: \"cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.390598 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8699\" (UniqueName: \"kubernetes.io/projected/d74f0516-3682-4b80-beb4-f7237c4daff8-kube-api-access-p8699\") pod \"designate-operator-controller-manager-6d9697b7f4-qnklz\" (UID: \"d74f0516-3682-4b80-beb4-f7237c4daff8\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.390642 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkmf7\" (UniqueName: \"kubernetes.io/projected/63c547ee-bb8c-4f1a-b29a-12a80c1749f4-kube-api-access-kkmf7\") pod \"heat-operator-controller-manager-69d6db494d-564mj\" (UID: \"63c547ee-bb8c-4f1a-b29a-12a80c1749f4\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.390666 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.390690 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlr7b\" (UniqueName: \"kubernetes.io/projected/ad043c5b-465c-43b2-b73c-45e605fc7b25-kube-api-access-vlr7b\") pod \"ironic-operator-controller-manager-5f4b8bd54d-zhlxl\" (UID: \"ad043c5b-465c-43b2-b73c-45e605fc7b25\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.390736 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f9xl\" (UniqueName: \"kubernetes.io/projected/6dd1801b-ac57-4d28-9554-ef8ad7148c08-kube-api-access-7f9xl\") pod \"glance-operator-controller-manager-8886f4c47-pgdzr\" (UID: \"6dd1801b-ac57-4d28-9554-ef8ad7148c08\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.390778 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw4zc\" (UniqueName: \"kubernetes.io/projected/d9f68e89-0c29-4d3d-9dae-7448ab35206f-kube-api-access-sw4zc\") pod \"horizon-operator-controller-manager-5fb775575f-hvnx5\" (UID: \"d9f68e89-0c29-4d3d-9dae-7448ab35206f\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.391028 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65kzb\" (UniqueName: \"kubernetes.io/projected/a6153ce6-006d-4351-8a9b-aa0abe1cbdad-kube-api-access-65kzb\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-6pbkt\" (UID: \"a6153ce6-006d-4351-8a9b-aa0abe1cbdad\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.391063 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfvhk\" (UniqueName: \"kubernetes.io/projected/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-kube-api-access-sfvhk\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.415849 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f9xl\" (UniqueName: \"kubernetes.io/projected/6dd1801b-ac57-4d28-9554-ef8ad7148c08-kube-api-access-7f9xl\") pod \"glance-operator-controller-manager-8886f4c47-pgdzr\" (UID: \"6dd1801b-ac57-4d28-9554-ef8ad7148c08\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.416534 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.417426 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.421401 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2w6sd" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.421660 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.422651 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.425614 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-snngj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.426420 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65kzb\" (UniqueName: \"kubernetes.io/projected/a6153ce6-006d-4351-8a9b-aa0abe1cbdad-kube-api-access-65kzb\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-6pbkt\" (UID: \"a6153ce6-006d-4351-8a9b-aa0abe1cbdad\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.433027 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.438387 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4wrv\" (UniqueName: \"kubernetes.io/projected/cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce-kube-api-access-w4wrv\") pod \"cinder-operator-controller-manager-8d874c8fc-dplk2\" (UID: \"cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.451368 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.452173 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.454930 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-mqqkb" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.461359 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8699\" (UniqueName: \"kubernetes.io/projected/d74f0516-3682-4b80-beb4-f7237c4daff8-kube-api-access-p8699\") pod \"designate-operator-controller-manager-6d9697b7f4-qnklz\" (UID: \"d74f0516-3682-4b80-beb4-f7237c4daff8\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.462760 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.478321 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.491964 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.492755 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.492969 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.493086 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkmf7\" (UniqueName: \"kubernetes.io/projected/63c547ee-bb8c-4f1a-b29a-12a80c1749f4-kube-api-access-kkmf7\") pod \"heat-operator-controller-manager-69d6db494d-564mj\" (UID: \"63c547ee-bb8c-4f1a-b29a-12a80c1749f4\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.493188 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlr7b\" (UniqueName: \"kubernetes.io/projected/ad043c5b-465c-43b2-b73c-45e605fc7b25-kube-api-access-vlr7b\") pod \"ironic-operator-controller-manager-5f4b8bd54d-zhlxl\" (UID: \"ad043c5b-465c-43b2-b73c-45e605fc7b25\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.493349 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw4zc\" (UniqueName: \"kubernetes.io/projected/d9f68e89-0c29-4d3d-9dae-7448ab35206f-kube-api-access-sw4zc\") pod \"horizon-operator-controller-manager-5fb775575f-hvnx5\" (UID: \"d9f68e89-0c29-4d3d-9dae-7448ab35206f\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" Feb 03 09:25:28 crc kubenswrapper[4756]: E0203 09:25:28.493118 4756 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:28 crc kubenswrapper[4756]: E0203 09:25:28.493606 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert podName:aae19bbd-e8ec-4a6f-8e77-9dd46373e31b nodeName:}" failed. No retries permitted until 2026-02-03 09:25:28.993583287 +0000 UTC m=+920.144050672 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert") pod "infra-operator-controller-manager-79955696d6-vvkmq" (UID: "aae19bbd-e8ec-4a6f-8e77-9dd46373e31b") : secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.493488 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j56sk\" (UniqueName: \"kubernetes.io/projected/919fda51-9a13-4233-ba62-a85ac3559a02-kube-api-access-j56sk\") pod \"keystone-operator-controller-manager-84f48565d4-w7vdf\" (UID: \"919fda51-9a13-4233-ba62-a85ac3559a02\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.493797 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b7pk\" (UniqueName: \"kubernetes.io/projected/aacb5f03-cd60-4589-9758-9fb4e1b68030-kube-api-access-7b7pk\") pod \"mariadb-operator-controller-manager-67bf948998-5q5s7\" (UID: \"aacb5f03-cd60-4589-9758-9fb4e1b68030\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.493915 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfvhk\" (UniqueName: \"kubernetes.io/projected/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-kube-api-access-sfvhk\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.494080 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq6jq\" (UniqueName: \"kubernetes.io/projected/d8ad4baf-2108-4286-a2f6-5a8ab537185c-kube-api-access-xq6jq\") pod \"manila-operator-controller-manager-7dd968899f-jwlpw\" (UID: \"d8ad4baf-2108-4286-a2f6-5a8ab537185c\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.518151 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-sdft2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.535324 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.536991 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlr7b\" (UniqueName: \"kubernetes.io/projected/ad043c5b-465c-43b2-b73c-45e605fc7b25-kube-api-access-vlr7b\") pod \"ironic-operator-controller-manager-5f4b8bd54d-zhlxl\" (UID: \"ad043c5b-465c-43b2-b73c-45e605fc7b25\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.543564 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.544570 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.547129 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw4zc\" (UniqueName: \"kubernetes.io/projected/d9f68e89-0c29-4d3d-9dae-7448ab35206f-kube-api-access-sw4zc\") pod \"horizon-operator-controller-manager-5fb775575f-hvnx5\" (UID: \"d9f68e89-0c29-4d3d-9dae-7448ab35206f\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.547328 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-twcgq" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.547615 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.547852 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.548491 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.559552 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.559879 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.561219 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfvhk\" (UniqueName: \"kubernetes.io/projected/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-kube-api-access-sfvhk\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.561646 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-j5fw9" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.589422 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.589711 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkmf7\" (UniqueName: \"kubernetes.io/projected/63c547ee-bb8c-4f1a-b29a-12a80c1749f4-kube-api-access-kkmf7\") pod \"heat-operator-controller-manager-69d6db494d-564mj\" (UID: \"63c547ee-bb8c-4f1a-b29a-12a80c1749f4\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.596733 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq6jq\" (UniqueName: \"kubernetes.io/projected/d8ad4baf-2108-4286-a2f6-5a8ab537185c-kube-api-access-xq6jq\") pod \"manila-operator-controller-manager-7dd968899f-jwlpw\" (UID: \"d8ad4baf-2108-4286-a2f6-5a8ab537185c\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.596852 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j56sk\" (UniqueName: \"kubernetes.io/projected/919fda51-9a13-4233-ba62-a85ac3559a02-kube-api-access-j56sk\") pod \"keystone-operator-controller-manager-84f48565d4-w7vdf\" (UID: \"919fda51-9a13-4233-ba62-a85ac3559a02\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.596889 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b7pk\" (UniqueName: \"kubernetes.io/projected/aacb5f03-cd60-4589-9758-9fb4e1b68030-kube-api-access-7b7pk\") pod \"mariadb-operator-controller-manager-67bf948998-5q5s7\" (UID: \"aacb5f03-cd60-4589-9758-9fb4e1b68030\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.604879 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.613716 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.625846 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b7pk\" (UniqueName: \"kubernetes.io/projected/aacb5f03-cd60-4589-9758-9fb4e1b68030-kube-api-access-7b7pk\") pod \"mariadb-operator-controller-manager-67bf948998-5q5s7\" (UID: \"aacb5f03-cd60-4589-9758-9fb4e1b68030\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.631223 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq6jq\" (UniqueName: \"kubernetes.io/projected/d8ad4baf-2108-4286-a2f6-5a8ab537185c-kube-api-access-xq6jq\") pod \"manila-operator-controller-manager-7dd968899f-jwlpw\" (UID: \"d8ad4baf-2108-4286-a2f6-5a8ab537185c\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.644116 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.655844 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.673405 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j56sk\" (UniqueName: \"kubernetes.io/projected/919fda51-9a13-4233-ba62-a85ac3559a02-kube-api-access-j56sk\") pod \"keystone-operator-controller-manager-84f48565d4-w7vdf\" (UID: \"919fda51-9a13-4233-ba62-a85ac3559a02\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.679932 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.681356 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.684471 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.684696 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9gt5p" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.699023 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.701174 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.702311 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.705968 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-ttlrw" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.706600 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjq7\" (UniqueName: \"kubernetes.io/projected/999c5dd9-6dbb-44b9-bdd7-52c146d34e87-kube-api-access-2tjq7\") pod \"nova-operator-controller-manager-55bff696bd-qnwj2\" (UID: \"999c5dd9-6dbb-44b9-bdd7-52c146d34e87\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.706655 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9g9l\" (UniqueName: \"kubernetes.io/projected/932503a4-3932-4127-a2b2-a2e0291782b7-kube-api-access-z9g9l\") pod \"octavia-operator-controller-manager-6687f8d877-bqthj\" (UID: \"932503a4-3932-4127-a2b2-a2e0291782b7\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.706712 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59vls\" (UniqueName: \"kubernetes.io/projected/4483922a-5bb6-40da-bb78-cdeca488ccd8-kube-api-access-59vls\") pod \"neutron-operator-controller-manager-585dbc889-gn2d2\" (UID: \"4483922a-5bb6-40da-bb78-cdeca488ccd8\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.712515 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.713753 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.717760 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-mttvb" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.725345 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.743891 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.763776 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.764915 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.767536 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-gx4lg" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.771999 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.802857 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.803831 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.808291 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9b9k\" (UniqueName: \"kubernetes.io/projected/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-kube-api-access-s9b9k\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.808346 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9g9l\" (UniqueName: \"kubernetes.io/projected/932503a4-3932-4127-a2b2-a2e0291782b7-kube-api-access-z9g9l\") pod \"octavia-operator-controller-manager-6687f8d877-bqthj\" (UID: \"932503a4-3932-4127-a2b2-a2e0291782b7\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.808425 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59vls\" (UniqueName: \"kubernetes.io/projected/4483922a-5bb6-40da-bb78-cdeca488ccd8-kube-api-access-59vls\") pod \"neutron-operator-controller-manager-585dbc889-gn2d2\" (UID: \"4483922a-5bb6-40da-bb78-cdeca488ccd8\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.808466 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.808504 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grxm9\" (UniqueName: \"kubernetes.io/projected/bd1225c5-fbb5-40f5-8e71-6c1075feeb16-kube-api-access-grxm9\") pod \"ovn-operator-controller-manager-788c46999f-wl9ft\" (UID: \"bd1225c5-fbb5-40f5-8e71-6c1075feeb16\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.808545 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tjq7\" (UniqueName: \"kubernetes.io/projected/999c5dd9-6dbb-44b9-bdd7-52c146d34e87-kube-api-access-2tjq7\") pod \"nova-operator-controller-manager-55bff696bd-qnwj2\" (UID: \"999c5dd9-6dbb-44b9-bdd7-52c146d34e87\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.812882 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.827510 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.827798 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.828500 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.835975 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-v97rn" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.842161 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59vls\" (UniqueName: \"kubernetes.io/projected/4483922a-5bb6-40da-bb78-cdeca488ccd8-kube-api-access-59vls\") pod \"neutron-operator-controller-manager-585dbc889-gn2d2\" (UID: \"4483922a-5bb6-40da-bb78-cdeca488ccd8\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.842797 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.854177 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tjq7\" (UniqueName: \"kubernetes.io/projected/999c5dd9-6dbb-44b9-bdd7-52c146d34e87-kube-api-access-2tjq7\") pod \"nova-operator-controller-manager-55bff696bd-qnwj2\" (UID: \"999c5dd9-6dbb-44b9-bdd7-52c146d34e87\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.854806 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9g9l\" (UniqueName: \"kubernetes.io/projected/932503a4-3932-4127-a2b2-a2e0291782b7-kube-api-access-z9g9l\") pod \"octavia-operator-controller-manager-6687f8d877-bqthj\" (UID: \"932503a4-3932-4127-a2b2-a2e0291782b7\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.872908 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.906512 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.908196 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.922809 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdccx\" (UniqueName: \"kubernetes.io/projected/58c01e66-fe0e-45f2-8caf-df41e1dad607-kube-api-access-tdccx\") pod \"placement-operator-controller-manager-5b964cf4cd-7rrf6\" (UID: \"58c01e66-fe0e-45f2-8caf-df41e1dad607\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.922953 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.922993 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9btm\" (UniqueName: \"kubernetes.io/projected/51a902ff-23e6-4578-809f-0de9a6c5cbae-kube-api-access-m9btm\") pod \"swift-operator-controller-manager-68fc8c869-w87dt\" (UID: \"51a902ff-23e6-4578-809f-0de9a6c5cbae\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.923021 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw8fw\" (UniqueName: \"kubernetes.io/projected/04edde9e-f443-4e9e-929b-65e2d9258ff2-kube-api-access-gw8fw\") pod \"test-operator-controller-manager-56f8bfcd9f-l5wms\" (UID: \"04edde9e-f443-4e9e-929b-65e2d9258ff2\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.923074 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grxm9\" (UniqueName: \"kubernetes.io/projected/bd1225c5-fbb5-40f5-8e71-6c1075feeb16-kube-api-access-grxm9\") pod \"ovn-operator-controller-manager-788c46999f-wl9ft\" (UID: \"bd1225c5-fbb5-40f5-8e71-6c1075feeb16\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.923168 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rd8g\" (UniqueName: \"kubernetes.io/projected/d7f70a3a-d4db-4d12-8838-9d8923f06213-kube-api-access-6rd8g\") pod \"telemetry-operator-controller-manager-64b5b76f97-2cw2v\" (UID: \"d7f70a3a-d4db-4d12-8838-9d8923f06213\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.923218 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9b9k\" (UniqueName: \"kubernetes.io/projected/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-kube-api-access-s9b9k\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:28 crc kubenswrapper[4756]: E0203 09:25:28.923966 4756 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:28 crc kubenswrapper[4756]: E0203 09:25:28.924023 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert podName:a0ebd520-4c15-4c12-89fe-2eb8a407ca25 nodeName:}" failed. No retries permitted until 2026-02-03 09:25:29.424003794 +0000 UTC m=+920.574471169 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" (UID: "a0ebd520-4c15-4c12-89fe-2eb8a407ca25") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.923974 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-nsltg" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.926629 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.927967 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.946584 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.971711 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grxm9\" (UniqueName: \"kubernetes.io/projected/bd1225c5-fbb5-40f5-8e71-6c1075feeb16-kube-api-access-grxm9\") pod \"ovn-operator-controller-manager-788c46999f-wl9ft\" (UID: \"bd1225c5-fbb5-40f5-8e71-6c1075feeb16\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.991798 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-9rdzw"] Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.992895 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" Feb 03 09:25:28 crc kubenswrapper[4756]: I0203 09:25:28.992897 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9b9k\" (UniqueName: \"kubernetes.io/projected/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-kube-api-access-s9b9k\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.003988 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-n9482" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.007471 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-9rdzw"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.024434 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9btm\" (UniqueName: \"kubernetes.io/projected/51a902ff-23e6-4578-809f-0de9a6c5cbae-kube-api-access-m9btm\") pod \"swift-operator-controller-manager-68fc8c869-w87dt\" (UID: \"51a902ff-23e6-4578-809f-0de9a6c5cbae\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.034627 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw8fw\" (UniqueName: \"kubernetes.io/projected/04edde9e-f443-4e9e-929b-65e2d9258ff2-kube-api-access-gw8fw\") pod \"test-operator-controller-manager-56f8bfcd9f-l5wms\" (UID: \"04edde9e-f443-4e9e-929b-65e2d9258ff2\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.034812 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rd8g\" (UniqueName: \"kubernetes.io/projected/d7f70a3a-d4db-4d12-8838-9d8923f06213-kube-api-access-6rd8g\") pod \"telemetry-operator-controller-manager-64b5b76f97-2cw2v\" (UID: \"d7f70a3a-d4db-4d12-8838-9d8923f06213\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.034867 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgw7d\" (UniqueName: \"kubernetes.io/projected/ef6310fb-d44c-449f-b3d0-72c20e4b3d7b-kube-api-access-xgw7d\") pod \"watcher-operator-controller-manager-564965969-9rdzw\" (UID: \"ef6310fb-d44c-449f-b3d0-72c20e4b3d7b\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.034926 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdccx\" (UniqueName: \"kubernetes.io/projected/58c01e66-fe0e-45f2-8caf-df41e1dad607-kube-api-access-tdccx\") pod \"placement-operator-controller-manager-5b964cf4cd-7rrf6\" (UID: \"58c01e66-fe0e-45f2-8caf-df41e1dad607\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.034977 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.035151 4756 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.035220 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert podName:aae19bbd-e8ec-4a6f-8e77-9dd46373e31b nodeName:}" failed. No retries permitted until 2026-02-03 09:25:30.035201409 +0000 UTC m=+921.185668784 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert") pod "infra-operator-controller-manager-79955696d6-vvkmq" (UID: "aae19bbd-e8ec-4a6f-8e77-9dd46373e31b") : secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.043409 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.052298 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ftt9v"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.068554 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.068579 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9btm\" (UniqueName: \"kubernetes.io/projected/51a902ff-23e6-4578-809f-0de9a6c5cbae-kube-api-access-m9btm\") pod \"swift-operator-controller-manager-68fc8c869-w87dt\" (UID: \"51a902ff-23e6-4578-809f-0de9a6c5cbae\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.069423 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.073728 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.073952 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.074773 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-glm2s" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.084277 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.093133 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rd8g\" (UniqueName: \"kubernetes.io/projected/d7f70a3a-d4db-4d12-8838-9d8923f06213-kube-api-access-6rd8g\") pod \"telemetry-operator-controller-manager-64b5b76f97-2cw2v\" (UID: \"d7f70a3a-d4db-4d12-8838-9d8923f06213\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.104712 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.105790 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.107738 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8gqxf" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.108791 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.118568 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw8fw\" (UniqueName: \"kubernetes.io/projected/04edde9e-f443-4e9e-929b-65e2d9258ff2-kube-api-access-gw8fw\") pod \"test-operator-controller-manager-56f8bfcd9f-l5wms\" (UID: \"04edde9e-f443-4e9e-929b-65e2d9258ff2\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.119224 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdccx\" (UniqueName: \"kubernetes.io/projected/58c01e66-fe0e-45f2-8caf-df41e1dad607-kube-api-access-tdccx\") pod \"placement-operator-controller-manager-5b964cf4cd-7rrf6\" (UID: \"58c01e66-fe0e-45f2-8caf-df41e1dad607\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.140274 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgw7d\" (UniqueName: \"kubernetes.io/projected/ef6310fb-d44c-449f-b3d0-72c20e4b3d7b-kube-api-access-xgw7d\") pod \"watcher-operator-controller-manager-564965969-9rdzw\" (UID: \"ef6310fb-d44c-449f-b3d0-72c20e4b3d7b\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.140867 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.151574 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.156073 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.158203 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgw7d\" (UniqueName: \"kubernetes.io/projected/ef6310fb-d44c-449f-b3d0-72c20e4b3d7b-kube-api-access-xgw7d\") pod \"watcher-operator-controller-manager-564965969-9rdzw\" (UID: \"ef6310fb-d44c-449f-b3d0-72c20e4b3d7b\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.242196 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhzcg\" (UniqueName: \"kubernetes.io/projected/9fe0931c-fae1-4dfd-968d-bccf61682fa5-kube-api-access-nhzcg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lqghx\" (UID: \"9fe0931c-fae1-4dfd-968d-bccf61682fa5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.242286 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2pjx\" (UniqueName: \"kubernetes.io/projected/2183ff26-c37d-472e-bf53-e0beba79a31d-kube-api-access-s2pjx\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.242317 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.242425 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.256118 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.312274 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.330293 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.343770 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhzcg\" (UniqueName: \"kubernetes.io/projected/9fe0931c-fae1-4dfd-968d-bccf61682fa5-kube-api-access-nhzcg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lqghx\" (UID: \"9fe0931c-fae1-4dfd-968d-bccf61682fa5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.343855 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2pjx\" (UniqueName: \"kubernetes.io/projected/2183ff26-c37d-472e-bf53-e0beba79a31d-kube-api-access-s2pjx\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.343885 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.343994 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.344139 4756 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.344207 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:29.844187222 +0000 UTC m=+920.994654597 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "webhook-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.345183 4756 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.345235 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:29.845222014 +0000 UTC m=+920.995689389 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "metrics-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.365252 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.378030 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhzcg\" (UniqueName: \"kubernetes.io/projected/9fe0931c-fae1-4dfd-968d-bccf61682fa5-kube-api-access-nhzcg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lqghx\" (UID: \"9fe0931c-fae1-4dfd-968d-bccf61682fa5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.379241 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2pjx\" (UniqueName: \"kubernetes.io/projected/2183ff26-c37d-472e-bf53-e0beba79a31d-kube-api-access-s2pjx\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.445339 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.445552 4756 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.445665 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert podName:a0ebd520-4c15-4c12-89fe-2eb8a407ca25 nodeName:}" failed. No retries permitted until 2026-02-03 09:25:30.445640722 +0000 UTC m=+921.596108097 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" (UID: "a0ebd520-4c15-4c12-89fe-2eb8a407ca25") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.490758 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.543177 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.797048 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" event={"ID":"a6153ce6-006d-4351-8a9b-aa0abe1cbdad","Type":"ContainerStarted","Data":"78004fe1faab8db08255ecd501bcd2bb285f06f9531f4de83c3766712ca0fb14"} Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.799856 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" event={"ID":"d74f0516-3682-4b80-beb4-f7237c4daff8","Type":"ContainerStarted","Data":"495e8c0e5b6f13995eece40b7c4f2d6f1b217e4446e0298ce3de8799e5567ea6"} Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.846449 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr"] Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.859308 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5"] Feb 03 09:25:29 crc kubenswrapper[4756]: W0203 09:25:29.859781 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9f68e89_0c29_4d3d_9dae_7448ab35206f.slice/crio-2e84f6e6d455c17d014dd637711c93f3a20023e497583969d48be2580c3a27c8 WatchSource:0}: Error finding container 2e84f6e6d455c17d014dd637711c93f3a20023e497583969d48be2580c3a27c8: Status 404 returned error can't find the container with id 2e84f6e6d455c17d014dd637711c93f3a20023e497583969d48be2580c3a27c8 Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.865239 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.865307 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.865409 4756 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.865447 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:30.865432447 +0000 UTC m=+922.015899812 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "metrics-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.865542 4756 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: E0203 09:25:29.865578 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:30.865568621 +0000 UTC m=+922.016035996 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "webhook-server-cert" not found Feb 03 09:25:29 crc kubenswrapper[4756]: I0203 09:25:29.883728 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-564mj"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.069380 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.069589 4756 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.069694 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert podName:aae19bbd-e8ec-4a6f-8e77-9dd46373e31b nodeName:}" failed. No retries permitted until 2026-02-03 09:25:32.069657537 +0000 UTC m=+923.220124902 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert") pod "infra-operator-controller-manager-79955696d6-vvkmq" (UID: "aae19bbd-e8ec-4a6f-8e77-9dd46373e31b") : secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.332606 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.346732 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.374036 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft"] Feb 03 09:25:30 crc kubenswrapper[4756]: W0203 09:25:30.379349 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaacb5f03_cd60_4589_9758_9fb4e1b68030.slice/crio-57db38a38005a14ad270c6fdcd2f9b86721bee9d93f95b6fc2921862b49ec6ef WatchSource:0}: Error finding container 57db38a38005a14ad270c6fdcd2f9b86721bee9d93f95b6fc2921862b49ec6ef: Status 404 returned error can't find the container with id 57db38a38005a14ad270c6fdcd2f9b86721bee9d93f95b6fc2921862b49ec6ef Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.386249 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.404863 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.415044 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7"] Feb 03 09:25:30 crc kubenswrapper[4756]: W0203 09:25:30.441951 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fe0931c_fae1_4dfd_968d_bccf61682fa5.slice/crio-1392fa69ac8c6b4566933cf994fe6974bd0f76bd2eeb9d950c269686784fc892 WatchSource:0}: Error finding container 1392fa69ac8c6b4566933cf994fe6974bd0f76bd2eeb9d950c269686784fc892: Status 404 returned error can't find the container with id 1392fa69ac8c6b4566933cf994fe6974bd0f76bd2eeb9d950c269686784fc892 Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.450021 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nhzcg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-lqghx_openstack-operators(9fe0931c-fae1-4dfd-968d-bccf61682fa5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.452497 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" podUID="9fe0931c-fae1-4dfd-968d-bccf61682fa5" Feb 03 09:25:30 crc kubenswrapper[4756]: W0203 09:25:30.457507 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef6310fb_d44c_449f_b3d0_72c20e4b3d7b.slice/crio-6313c44c189bf2fc13c94bc6fa14e8eb4d2f10a70332ae5971a2851b65df5b93 WatchSource:0}: Error finding container 6313c44c189bf2fc13c94bc6fa14e8eb4d2f10a70332ae5971a2851b65df5b93: Status 404 returned error can't find the container with id 6313c44c189bf2fc13c94bc6fa14e8eb4d2f10a70332ae5971a2851b65df5b93 Feb 03 09:25:30 crc kubenswrapper[4756]: W0203 09:25:30.457743 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51a902ff_23e6_4578_809f_0de9a6c5cbae.slice/crio-8ee115c6abcbdadcff946541c6b4e4819a838a5f0ab27fefc1d3ca02a320e4f2 WatchSource:0}: Error finding container 8ee115c6abcbdadcff946541c6b4e4819a838a5f0ab27fefc1d3ca02a320e4f2: Status 404 returned error can't find the container with id 8ee115c6abcbdadcff946541c6b4e4819a838a5f0ab27fefc1d3ca02a320e4f2 Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.474554 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.474722 4756 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.474773 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert podName:a0ebd520-4c15-4c12-89fe-2eb8a407ca25 nodeName:}" failed. No retries permitted until 2026-02-03 09:25:32.474759033 +0000 UTC m=+923.625226408 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" (UID: "a0ebd520-4c15-4c12-89fe-2eb8a407ca25") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.478410 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-9rdzw"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.499839 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl"] Feb 03 09:25:30 crc kubenswrapper[4756]: W0203 09:25:30.500799 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58c01e66_fe0e_45f2_8caf_df41e1dad607.slice/crio-d4c07a834fd029564474a362f749a74e99241722a38dc5188badf2ede6d42129 WatchSource:0}: Error finding container d4c07a834fd029564474a362f749a74e99241722a38dc5188badf2ede6d42129: Status 404 returned error can't find the container with id d4c07a834fd029564474a362f749a74e99241722a38dc5188badf2ede6d42129 Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.501280 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:42ad717de1b82267d244b016e5491a5b66a5c3deb6b8c2906a379e1296a2c382,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m9btm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68fc8c869-w87dt_openstack-operators(51a902ff-23e6-4578-809f-0de9a6c5cbae): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.503619 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" podUID="51a902ff-23e6-4578-809f-0de9a6c5cbae" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.506231 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:bead175f27e5f074f723694f3b66e5aa7238411bf8a27a267b9a2936e4465521,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vlr7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-5f4b8bd54d-zhlxl_openstack-operators(ad043c5b-465c-43b2-b73c-45e605fc7b25): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.506407 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tdccx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b964cf4cd-7rrf6_openstack-operators(58c01e66-fe0e-45f2-8caf-df41e1dad607): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.506554 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:6e21a1dda86ba365817102d23a5d4d2d5dcd1c4d8e5f8d74bd24548aa8c63898,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w4wrv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-8d874c8fc-dplk2_openstack-operators(cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.507356 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" podUID="ad043c5b-465c-43b2-b73c-45e605fc7b25" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.507545 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" podUID="58c01e66-fe0e-45f2-8caf-df41e1dad607" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.507615 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" podUID="cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce" Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.512640 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.534035 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.554567 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.561685 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.571765 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.585913 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.595407 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx"] Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.818161 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" event={"ID":"d9f68e89-0c29-4d3d-9dae-7448ab35206f","Type":"ContainerStarted","Data":"2e84f6e6d455c17d014dd637711c93f3a20023e497583969d48be2580c3a27c8"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.820635 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" event={"ID":"aacb5f03-cd60-4589-9758-9fb4e1b68030","Type":"ContainerStarted","Data":"57db38a38005a14ad270c6fdcd2f9b86721bee9d93f95b6fc2921862b49ec6ef"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.823575 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" event={"ID":"9fe0931c-fae1-4dfd-968d-bccf61682fa5","Type":"ContainerStarted","Data":"1392fa69ac8c6b4566933cf994fe6974bd0f76bd2eeb9d950c269686784fc892"} Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.826333 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" podUID="9fe0931c-fae1-4dfd-968d-bccf61682fa5" Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.828125 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" event={"ID":"bd1225c5-fbb5-40f5-8e71-6c1075feeb16","Type":"ContainerStarted","Data":"b6dc5f677c1365824f15d00fc67e953c7f2ec28ef7797b394ba59bdebb8f618c"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.830464 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" event={"ID":"58c01e66-fe0e-45f2-8caf-df41e1dad607","Type":"ContainerStarted","Data":"d4c07a834fd029564474a362f749a74e99241722a38dc5188badf2ede6d42129"} Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.833516 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" podUID="58c01e66-fe0e-45f2-8caf-df41e1dad607" Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.833796 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" event={"ID":"ad043c5b-465c-43b2-b73c-45e605fc7b25","Type":"ContainerStarted","Data":"cbff8646b93e8b017203f8c2b5b6790a3656c047e879b40876d35bbc1dae258e"} Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.835886 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:bead175f27e5f074f723694f3b66e5aa7238411bf8a27a267b9a2936e4465521\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" podUID="ad043c5b-465c-43b2-b73c-45e605fc7b25" Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.836812 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" event={"ID":"63c547ee-bb8c-4f1a-b29a-12a80c1749f4","Type":"ContainerStarted","Data":"d8567c57ffacaeec22049e3e796c793ff054f813d02716691f25a803ca6fecdd"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.843102 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" event={"ID":"4483922a-5bb6-40da-bb78-cdeca488ccd8","Type":"ContainerStarted","Data":"09235b48ebb3f40fea1eb663922bd72af9cfdf87c5be051186c242198169c828"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.846547 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" event={"ID":"d7f70a3a-d4db-4d12-8838-9d8923f06213","Type":"ContainerStarted","Data":"171df29d5b53d348f53e01015395350c4deda97fab7825946a1932a5dc5d7248"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.849316 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" event={"ID":"999c5dd9-6dbb-44b9-bdd7-52c146d34e87","Type":"ContainerStarted","Data":"db66adee9e0c55199cc3602c80a1e77f940b117c2387aa8b0b7bdf31d405abd5"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.890916 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" event={"ID":"919fda51-9a13-4233-ba62-a85ac3559a02","Type":"ContainerStarted","Data":"85d72a3134de8df66db1868e8528577f8e98666451da8c0615bf9c17d625391f"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.891898 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.892705 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.892263 4756 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.893671 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:32.89365106 +0000 UTC m=+924.044118435 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "webhook-server-cert" not found Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.894397 4756 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.894528 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:32.894516127 +0000 UTC m=+924.044983502 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "metrics-server-cert" not found Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.905565 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" event={"ID":"d8ad4baf-2108-4286-a2f6-5a8ab537185c","Type":"ContainerStarted","Data":"0990e14f07ea8fc6043209ae860fddf78ccb862d79cf8dfab41c75e69e9c26cf"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.913357 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" event={"ID":"51a902ff-23e6-4578-809f-0de9a6c5cbae","Type":"ContainerStarted","Data":"8ee115c6abcbdadcff946541c6b4e4819a838a5f0ab27fefc1d3ca02a320e4f2"} Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.915715 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:42ad717de1b82267d244b016e5491a5b66a5c3deb6b8c2906a379e1296a2c382\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" podUID="51a902ff-23e6-4578-809f-0de9a6c5cbae" Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.916901 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" event={"ID":"932503a4-3932-4127-a2b2-a2e0291782b7","Type":"ContainerStarted","Data":"62657af53012b6bf92e768180483602b0cfe7da1bcd2d282702c49ebd59fc0a2"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.926781 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" event={"ID":"cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce","Type":"ContainerStarted","Data":"a2a139c091f842d5c36ff1ff41de2dde9fbd5205c94a2a466ab4cb6b28593cf7"} Feb 03 09:25:30 crc kubenswrapper[4756]: E0203 09:25:30.929015 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:6e21a1dda86ba365817102d23a5d4d2d5dcd1c4d8e5f8d74bd24548aa8c63898\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" podUID="cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce" Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.929799 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" event={"ID":"ef6310fb-d44c-449f-b3d0-72c20e4b3d7b","Type":"ContainerStarted","Data":"6313c44c189bf2fc13c94bc6fa14e8eb4d2f10a70332ae5971a2851b65df5b93"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.931344 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" event={"ID":"04edde9e-f443-4e9e-929b-65e2d9258ff2","Type":"ContainerStarted","Data":"f10374015017cdffce635a9f901d0069764ed33194054a84df60a55104cd37ed"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.932980 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" event={"ID":"6dd1801b-ac57-4d28-9554-ef8ad7148c08","Type":"ContainerStarted","Data":"9d337b68e0ca1ff84b10948102bbeeb00983f3ad77d0595f64d0603ccf9703a5"} Feb 03 09:25:30 crc kubenswrapper[4756]: I0203 09:25:30.933231 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ftt9v" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerName="registry-server" containerID="cri-o://8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300" gracePeriod=2 Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.606265 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.616578 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-catalog-content\") pod \"64eb5569-95ab-4135-9d4f-4bda9904b174\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.616644 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-utilities\") pod \"64eb5569-95ab-4135-9d4f-4bda9904b174\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.616723 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95gw8\" (UniqueName: \"kubernetes.io/projected/64eb5569-95ab-4135-9d4f-4bda9904b174-kube-api-access-95gw8\") pod \"64eb5569-95ab-4135-9d4f-4bda9904b174\" (UID: \"64eb5569-95ab-4135-9d4f-4bda9904b174\") " Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.617670 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-utilities" (OuterVolumeSpecName: "utilities") pod "64eb5569-95ab-4135-9d4f-4bda9904b174" (UID: "64eb5569-95ab-4135-9d4f-4bda9904b174"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.652751 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64eb5569-95ab-4135-9d4f-4bda9904b174-kube-api-access-95gw8" (OuterVolumeSpecName: "kube-api-access-95gw8") pod "64eb5569-95ab-4135-9d4f-4bda9904b174" (UID: "64eb5569-95ab-4135-9d4f-4bda9904b174"). InnerVolumeSpecName "kube-api-access-95gw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.700399 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64eb5569-95ab-4135-9d4f-4bda9904b174" (UID: "64eb5569-95ab-4135-9d4f-4bda9904b174"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.719091 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.719128 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64eb5569-95ab-4135-9d4f-4bda9904b174-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.719140 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95gw8\" (UniqueName: \"kubernetes.io/projected/64eb5569-95ab-4135-9d4f-4bda9904b174-kube-api-access-95gw8\") on node \"crc\" DevicePath \"\"" Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.946056 4756 generic.go:334] "Generic (PLEG): container finished" podID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerID="8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300" exitCode=0 Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.946496 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ftt9v" event={"ID":"64eb5569-95ab-4135-9d4f-4bda9904b174","Type":"ContainerDied","Data":"8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300"} Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.946550 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ftt9v" event={"ID":"64eb5569-95ab-4135-9d4f-4bda9904b174","Type":"ContainerDied","Data":"9f1b30c44f4eef43d1cf1c1e4b3b2bdf3a6a353c607c790c317a48a8df600971"} Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.946570 4756 scope.go:117] "RemoveContainer" containerID="8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300" Feb 03 09:25:31 crc kubenswrapper[4756]: I0203 09:25:31.946658 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ftt9v" Feb 03 09:25:31 crc kubenswrapper[4756]: E0203 09:25:31.949622 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:42ad717de1b82267d244b016e5491a5b66a5c3deb6b8c2906a379e1296a2c382\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" podUID="51a902ff-23e6-4578-809f-0de9a6c5cbae" Feb 03 09:25:31 crc kubenswrapper[4756]: E0203 09:25:31.949869 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:bead175f27e5f074f723694f3b66e5aa7238411bf8a27a267b9a2936e4465521\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" podUID="ad043c5b-465c-43b2-b73c-45e605fc7b25" Feb 03 09:25:31 crc kubenswrapper[4756]: E0203 09:25:31.950071 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" podUID="58c01e66-fe0e-45f2-8caf-df41e1dad607" Feb 03 09:25:31 crc kubenswrapper[4756]: E0203 09:25:31.950190 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" podUID="9fe0931c-fae1-4dfd-968d-bccf61682fa5" Feb 03 09:25:31 crc kubenswrapper[4756]: E0203 09:25:31.950540 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:6e21a1dda86ba365817102d23a5d4d2d5dcd1c4d8e5f8d74bd24548aa8c63898\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" podUID="cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.011648 4756 scope.go:117] "RemoveContainer" containerID="36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.095355 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ftt9v"] Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.102510 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ftt9v"] Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.108289 4756 scope.go:117] "RemoveContainer" containerID="9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.130159 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.130735 4756 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.130779 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert podName:aae19bbd-e8ec-4a6f-8e77-9dd46373e31b nodeName:}" failed. No retries permitted until 2026-02-03 09:25:36.1307655 +0000 UTC m=+927.281232875 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert") pod "infra-operator-controller-manager-79955696d6-vvkmq" (UID: "aae19bbd-e8ec-4a6f-8e77-9dd46373e31b") : secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.223114 4756 scope.go:117] "RemoveContainer" containerID="8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300" Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.233141 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300\": container with ID starting with 8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300 not found: ID does not exist" containerID="8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.233178 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300"} err="failed to get container status \"8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300\": rpc error: code = NotFound desc = could not find container \"8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300\": container with ID starting with 8487b52328a8ef5651eb9f7f6f5fe798de395492e7590b8fb602603149d81300 not found: ID does not exist" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.233201 4756 scope.go:117] "RemoveContainer" containerID="36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2" Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.236061 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2\": container with ID starting with 36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2 not found: ID does not exist" containerID="36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.236092 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2"} err="failed to get container status \"36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2\": rpc error: code = NotFound desc = could not find container \"36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2\": container with ID starting with 36fcb36aed5956d0d95f22b1c5d3417b0f55544e227628ee0f8de0239d15c4d2 not found: ID does not exist" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.236145 4756 scope.go:117] "RemoveContainer" containerID="9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e" Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.236404 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e\": container with ID starting with 9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e not found: ID does not exist" containerID="9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.236426 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e"} err="failed to get container status \"9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e\": rpc error: code = NotFound desc = could not find container \"9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e\": container with ID starting with 9fe6ee85a0969d2944d41ef6fc196131d18c98d70e8a07e8513d72250cfc1a1e not found: ID does not exist" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.543355 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.543564 4756 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.543734 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert podName:a0ebd520-4c15-4c12-89fe-2eb8a407ca25 nodeName:}" failed. No retries permitted until 2026-02-03 09:25:36.543715061 +0000 UTC m=+927.694182426 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" (UID: "a0ebd520-4c15-4c12-89fe-2eb8a407ca25") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.956853 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:32 crc kubenswrapper[4756]: I0203 09:25:32.956968 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.957067 4756 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.957164 4756 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.957186 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:36.957136657 +0000 UTC m=+928.107604032 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "webhook-server-cert" not found Feb 03 09:25:32 crc kubenswrapper[4756]: E0203 09:25:32.957235 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:36.9572199 +0000 UTC m=+928.107687275 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "metrics-server-cert" not found Feb 03 09:25:33 crc kubenswrapper[4756]: I0203 09:25:33.633904 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" path="/var/lib/kubelet/pods/64eb5569-95ab-4135-9d4f-4bda9904b174/volumes" Feb 03 09:25:36 crc kubenswrapper[4756]: I0203 09:25:36.216155 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:36 crc kubenswrapper[4756]: E0203 09:25:36.216718 4756 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:36 crc kubenswrapper[4756]: E0203 09:25:36.216764 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert podName:aae19bbd-e8ec-4a6f-8e77-9dd46373e31b nodeName:}" failed. No retries permitted until 2026-02-03 09:25:44.216751143 +0000 UTC m=+935.367218518 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert") pod "infra-operator-controller-manager-79955696d6-vvkmq" (UID: "aae19bbd-e8ec-4a6f-8e77-9dd46373e31b") : secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:36 crc kubenswrapper[4756]: I0203 09:25:36.623909 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:36 crc kubenswrapper[4756]: E0203 09:25:36.624166 4756 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:36 crc kubenswrapper[4756]: E0203 09:25:36.624222 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert podName:a0ebd520-4c15-4c12-89fe-2eb8a407ca25 nodeName:}" failed. No retries permitted until 2026-02-03 09:25:44.624208663 +0000 UTC m=+935.774676028 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" (UID: "a0ebd520-4c15-4c12-89fe-2eb8a407ca25") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:37 crc kubenswrapper[4756]: I0203 09:25:37.029800 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:37 crc kubenswrapper[4756]: E0203 09:25:37.029978 4756 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 03 09:25:37 crc kubenswrapper[4756]: E0203 09:25:37.030228 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:45.030210148 +0000 UTC m=+936.180677523 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "webhook-server-cert" not found Feb 03 09:25:37 crc kubenswrapper[4756]: I0203 09:25:37.030563 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:37 crc kubenswrapper[4756]: E0203 09:25:37.030686 4756 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 03 09:25:37 crc kubenswrapper[4756]: E0203 09:25:37.030721 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:25:45.030711364 +0000 UTC m=+936.181178739 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "metrics-server-cert" not found Feb 03 09:25:43 crc kubenswrapper[4756]: I0203 09:25:43.566542 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:25:43 crc kubenswrapper[4756]: I0203 09:25:43.567174 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:25:43 crc kubenswrapper[4756]: I0203 09:25:43.567231 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:25:43 crc kubenswrapper[4756]: I0203 09:25:43.567968 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef696d62114310b595359fdf0f823a6781a2df097280882ce3528c732940055c"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:25:43 crc kubenswrapper[4756]: I0203 09:25:43.568027 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://ef696d62114310b595359fdf0f823a6781a2df097280882ce3528c732940055c" gracePeriod=600 Feb 03 09:25:44 crc kubenswrapper[4756]: I0203 09:25:44.029645 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="ef696d62114310b595359fdf0f823a6781a2df097280882ce3528c732940055c" exitCode=0 Feb 03 09:25:44 crc kubenswrapper[4756]: I0203 09:25:44.029696 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"ef696d62114310b595359fdf0f823a6781a2df097280882ce3528c732940055c"} Feb 03 09:25:44 crc kubenswrapper[4756]: I0203 09:25:44.029734 4756 scope.go:117] "RemoveContainer" containerID="c20ecab788892a342dab2838ffafa76a4cbfbef0fe19794f920df87b66dcf0b1" Feb 03 09:25:44 crc kubenswrapper[4756]: E0203 09:25:44.189673 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:1f593e8d49d02b6484c89632192ae54771675c54fbd8426e3675b8e20ecfd7c4" Feb 03 09:25:44 crc kubenswrapper[4756]: E0203 09:25:44.189839 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:1f593e8d49d02b6484c89632192ae54771675c54fbd8426e3675b8e20ecfd7c4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7f9xl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-8886f4c47-pgdzr_openstack-operators(6dd1801b-ac57-4d28-9554-ef8ad7148c08): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:25:44 crc kubenswrapper[4756]: E0203 09:25:44.191027 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" podUID="6dd1801b-ac57-4d28-9554-ef8ad7148c08" Feb 03 09:25:44 crc kubenswrapper[4756]: I0203 09:25:44.254577 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:25:44 crc kubenswrapper[4756]: E0203 09:25:44.254737 4756 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:44 crc kubenswrapper[4756]: E0203 09:25:44.254824 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert podName:aae19bbd-e8ec-4a6f-8e77-9dd46373e31b nodeName:}" failed. No retries permitted until 2026-02-03 09:26:00.254806608 +0000 UTC m=+951.405273983 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert") pod "infra-operator-controller-manager-79955696d6-vvkmq" (UID: "aae19bbd-e8ec-4a6f-8e77-9dd46373e31b") : secret "infra-operator-webhook-server-cert" not found Feb 03 09:25:44 crc kubenswrapper[4756]: I0203 09:25:44.668513 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:25:44 crc kubenswrapper[4756]: E0203 09:25:44.669666 4756 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:44 crc kubenswrapper[4756]: E0203 09:25:44.669709 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert podName:a0ebd520-4c15-4c12-89fe-2eb8a407ca25 nodeName:}" failed. No retries permitted until 2026-02-03 09:26:00.66969442 +0000 UTC m=+951.820161795 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" (UID: "a0ebd520-4c15-4c12-89fe-2eb8a407ca25") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 03 09:25:45 crc kubenswrapper[4756]: E0203 09:25:45.036708 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:1f593e8d49d02b6484c89632192ae54771675c54fbd8426e3675b8e20ecfd7c4\\\"\"" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" podUID="6dd1801b-ac57-4d28-9554-ef8ad7148c08" Feb 03 09:25:45 crc kubenswrapper[4756]: I0203 09:25:45.074968 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:45 crc kubenswrapper[4756]: I0203 09:25:45.075056 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:25:45 crc kubenswrapper[4756]: E0203 09:25:45.075163 4756 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 03 09:25:45 crc kubenswrapper[4756]: E0203 09:25:45.075206 4756 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 03 09:25:45 crc kubenswrapper[4756]: E0203 09:25:45.075218 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:26:01.075198109 +0000 UTC m=+952.225665484 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "metrics-server-cert" not found Feb 03 09:25:45 crc kubenswrapper[4756]: E0203 09:25:45.075285 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs podName:2183ff26-c37d-472e-bf53-e0beba79a31d nodeName:}" failed. No retries permitted until 2026-02-03 09:26:01.075268001 +0000 UTC m=+952.225735376 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs") pod "openstack-operator-controller-manager-6c6ffddb96-rvk7w" (UID: "2183ff26-c37d-472e-bf53-e0beba79a31d") : secret "webhook-server-cert" not found Feb 03 09:25:45 crc kubenswrapper[4756]: E0203 09:25:45.523108 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:027cd7ab61ef5071d9ad6b729c95a98e51cd254642f01dc019d44cc98a9232f8" Feb 03 09:25:45 crc kubenswrapper[4756]: E0203 09:25:45.523645 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:027cd7ab61ef5071d9ad6b729c95a98e51cd254642f01dc019d44cc98a9232f8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sw4zc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5fb775575f-hvnx5_openstack-operators(d9f68e89-0c29-4d3d-9dae-7448ab35206f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:25:45 crc kubenswrapper[4756]: E0203 09:25:45.524864 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" podUID="d9f68e89-0c29-4d3d-9dae-7448ab35206f" Feb 03 09:25:46 crc kubenswrapper[4756]: E0203 09:25:46.051792 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:027cd7ab61ef5071d9ad6b729c95a98e51cd254642f01dc019d44cc98a9232f8\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" podUID="d9f68e89-0c29-4d3d-9dae-7448ab35206f" Feb 03 09:25:46 crc kubenswrapper[4756]: E0203 09:25:46.088493 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:5340b88039fac393da49ef4e181b2720c809c27a6bb30531a07a49342a1da45e" Feb 03 09:25:46 crc kubenswrapper[4756]: E0203 09:25:46.088680 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:5340b88039fac393da49ef4e181b2720c809c27a6bb30531a07a49342a1da45e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2tjq7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-55bff696bd-qnwj2_openstack-operators(999c5dd9-6dbb-44b9-bdd7-52c146d34e87): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:25:46 crc kubenswrapper[4756]: E0203 09:25:46.090050 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" podUID="999c5dd9-6dbb-44b9-bdd7-52c146d34e87" Feb 03 09:25:46 crc kubenswrapper[4756]: E0203 09:25:46.565395 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:319c969e88f109b26487a9f5a67203682803d7386424703ab7ca0340be99ae17" Feb 03 09:25:46 crc kubenswrapper[4756]: E0203 09:25:46.565653 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:319c969e88f109b26487a9f5a67203682803d7386424703ab7ca0340be99ae17,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j56sk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-84f48565d4-w7vdf_openstack-operators(919fda51-9a13-4233-ba62-a85ac3559a02): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:25:46 crc kubenswrapper[4756]: E0203 09:25:46.566881 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" podUID="919fda51-9a13-4233-ba62-a85ac3559a02" Feb 03 09:25:47 crc kubenswrapper[4756]: E0203 09:25:47.053936 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:5340b88039fac393da49ef4e181b2720c809c27a6bb30531a07a49342a1da45e\\\"\"" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" podUID="999c5dd9-6dbb-44b9-bdd7-52c146d34e87" Feb 03 09:25:47 crc kubenswrapper[4756]: E0203 09:25:47.054013 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:319c969e88f109b26487a9f5a67203682803d7386424703ab7ca0340be99ae17\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" podUID="919fda51-9a13-4233-ba62-a85ac3559a02" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.129938 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" event={"ID":"ef6310fb-d44c-449f-b3d0-72c20e4b3d7b","Type":"ContainerStarted","Data":"9ee294d8c75df6ba13a2d82ff46ce0e445b68dd03ded3501d158d41bfe10e661"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.130697 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.170236 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" event={"ID":"cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce","Type":"ContainerStarted","Data":"d919a53f1d702b2d11fbeb51de1d7beb0c757dd0032cb2f29f0a5be3176f4a14"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.170477 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.186602 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" event={"ID":"932503a4-3932-4127-a2b2-a2e0291782b7","Type":"ContainerStarted","Data":"188a25a8bf074db6c58b2a1d9b00866eb66c2656af56f2595e94fb38291bfcb0"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.187161 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.188395 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" podStartSLOduration=14.130400016 podStartE2EDuration="30.188376009s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.482839265 +0000 UTC m=+921.633306640" lastFinishedPulling="2026-02-03 09:25:46.540815258 +0000 UTC m=+937.691282633" observedRunningTime="2026-02-03 09:25:58.183508617 +0000 UTC m=+949.333976002" watchObservedRunningTime="2026-02-03 09:25:58.188376009 +0000 UTC m=+949.338843394" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.198056 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" event={"ID":"d8ad4baf-2108-4286-a2f6-5a8ab537185c","Type":"ContainerStarted","Data":"4df07e823a780c06e4791a699f4853e2faa1f2a1198de13ea2e80a906f972640"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.198327 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.211745 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" event={"ID":"ad043c5b-465c-43b2-b73c-45e605fc7b25","Type":"ContainerStarted","Data":"694d5a769dbc6223cf91e852ecb5b5ced969198de5c4300a42cad7bf4eba5bd2"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.212241 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.219575 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" podStartSLOduration=3.847273513 podStartE2EDuration="30.219557753s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.506477154 +0000 UTC m=+921.656944529" lastFinishedPulling="2026-02-03 09:25:56.878761384 +0000 UTC m=+948.029228769" observedRunningTime="2026-02-03 09:25:58.217065736 +0000 UTC m=+949.367533121" watchObservedRunningTime="2026-02-03 09:25:58.219557753 +0000 UTC m=+949.370025128" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.227048 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" event={"ID":"4483922a-5bb6-40da-bb78-cdeca488ccd8","Type":"ContainerStarted","Data":"dd4d2ac8c9465e904f2e2fe5c992b10eb075858693dc5ba4195d91217264da02"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.228159 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.235698 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" event={"ID":"aacb5f03-cd60-4589-9758-9fb4e1b68030","Type":"ContainerStarted","Data":"4101bf83196ad12c804cc17ecee96281b9e344a5791839d7423912ea521efa25"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.236354 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.258393 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" event={"ID":"d7f70a3a-d4db-4d12-8838-9d8923f06213","Type":"ContainerStarted","Data":"f0a6ea993c32d6c6eefc571b586a5ac8760bd949d4d8c059a2d9fb75c9bb8536"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.259021 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.260094 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" event={"ID":"63c547ee-bb8c-4f1a-b29a-12a80c1749f4","Type":"ContainerStarted","Data":"9f60ef5013dddd4c7e47f8d9e2b2cb6018a2029b478ca6519382af95e048dd51"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.260677 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.268499 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" event={"ID":"58c01e66-fe0e-45f2-8caf-df41e1dad607","Type":"ContainerStarted","Data":"43db9fc48dff02443fad58ff1c4e8bb61cbbea10eb91825b1636bdde9fb13853"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.269211 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.285205 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" event={"ID":"51a902ff-23e6-4578-809f-0de9a6c5cbae","Type":"ContainerStarted","Data":"b39488e5a7a727cf315e060a248be8f068e58dc09aefa14ad57c58e8c4231bbc"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.285921 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.288710 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" event={"ID":"d74f0516-3682-4b80-beb4-f7237c4daff8","Type":"ContainerStarted","Data":"37d799cf7a6480d8df272be8cb212b1228b76145ce9b2272e9d3a64bd010f326"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.289302 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.304574 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" event={"ID":"bd1225c5-fbb5-40f5-8e71-6c1075feeb16","Type":"ContainerStarted","Data":"aef2992ee5356162bbe757c9c71cf6d470411200359c91ca37a0bed6331d7fdb"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.304892 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" podStartSLOduration=14.170650553 podStartE2EDuration="30.304871078s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.404520068 +0000 UTC m=+921.554987443" lastFinishedPulling="2026-02-03 09:25:46.538740593 +0000 UTC m=+937.689207968" observedRunningTime="2026-02-03 09:25:58.248634262 +0000 UTC m=+949.399101637" watchObservedRunningTime="2026-02-03 09:25:58.304871078 +0000 UTC m=+949.455338453" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.305512 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.305671 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" podStartSLOduration=3.875022729 podStartE2EDuration="30.305664393s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.506102952 +0000 UTC m=+921.656570327" lastFinishedPulling="2026-02-03 09:25:56.936744626 +0000 UTC m=+948.087211991" observedRunningTime="2026-02-03 09:25:58.294168924 +0000 UTC m=+949.444636289" watchObservedRunningTime="2026-02-03 09:25:58.305664393 +0000 UTC m=+949.456131768" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.326604 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" event={"ID":"a6153ce6-006d-4351-8a9b-aa0abe1cbdad","Type":"ContainerStarted","Data":"a2cefda6bd87db610a3b49a439b80a7b4f34973b27444e9facb88f01fb62ad8d"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.326742 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.334381 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" event={"ID":"04edde9e-f443-4e9e-929b-65e2d9258ff2","Type":"ContainerStarted","Data":"35c699894e90c562b5d3a0ec34b4a36940ab474d463ae3caea591506737c2a92"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.335153 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.348560 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" podStartSLOduration=14.208150707 podStartE2EDuration="30.348544693s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.400906626 +0000 UTC m=+921.551374001" lastFinishedPulling="2026-02-03 09:25:46.541300612 +0000 UTC m=+937.691767987" observedRunningTime="2026-02-03 09:25:58.347703757 +0000 UTC m=+949.498171132" watchObservedRunningTime="2026-02-03 09:25:58.348544693 +0000 UTC m=+949.499012058" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.385082 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"0da902e4b7c741d00f45d02b231b9b641521e6c5b54725cc76fd568b0058caf7"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.395204 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" podStartSLOduration=13.531088974 podStartE2EDuration="30.39518742s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.380132747 +0000 UTC m=+921.530600122" lastFinishedPulling="2026-02-03 09:25:47.244231193 +0000 UTC m=+938.394698568" observedRunningTime="2026-02-03 09:25:58.39487362 +0000 UTC m=+949.545340995" watchObservedRunningTime="2026-02-03 09:25:58.39518742 +0000 UTC m=+949.545654795" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.397472 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" podStartSLOduration=13.248066061 podStartE2EDuration="30.397466551s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:29.393597376 +0000 UTC m=+920.544064751" lastFinishedPulling="2026-02-03 09:25:46.542997866 +0000 UTC m=+937.693465241" observedRunningTime="2026-02-03 09:25:58.365353268 +0000 UTC m=+949.515820643" watchObservedRunningTime="2026-02-03 09:25:58.397466551 +0000 UTC m=+949.547933926" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.407649 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" event={"ID":"9fe0931c-fae1-4dfd-968d-bccf61682fa5","Type":"ContainerStarted","Data":"e5309608bfb7f8cf272fb4bad5b0324d4880b095d8c65402e92646098c9bf0d8"} Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.426950 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" podStartSLOduration=13.570836875 podStartE2EDuration="30.426934892s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.388503158 +0000 UTC m=+921.538970533" lastFinishedPulling="2026-02-03 09:25:47.244601175 +0000 UTC m=+938.395068550" observedRunningTime="2026-02-03 09:25:58.423785804 +0000 UTC m=+949.574253179" watchObservedRunningTime="2026-02-03 09:25:58.426934892 +0000 UTC m=+949.577402267" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.455541 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" podStartSLOduration=3.971129193 podStartE2EDuration="30.455524646s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.506308129 +0000 UTC m=+921.656775504" lastFinishedPulling="2026-02-03 09:25:56.990703582 +0000 UTC m=+948.141170957" observedRunningTime="2026-02-03 09:25:58.452053197 +0000 UTC m=+949.602520572" watchObservedRunningTime="2026-02-03 09:25:58.455524646 +0000 UTC m=+949.605992021" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.516428 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" podStartSLOduration=4.678940166 podStartE2EDuration="30.516415008s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.501111206 +0000 UTC m=+921.651578581" lastFinishedPulling="2026-02-03 09:25:56.338586048 +0000 UTC m=+947.489053423" observedRunningTime="2026-02-03 09:25:58.481029243 +0000 UTC m=+949.631496618" watchObservedRunningTime="2026-02-03 09:25:58.516415008 +0000 UTC m=+949.666882383" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.518965 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" podStartSLOduration=13.865502141 podStartE2EDuration="30.518957587s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:29.888709834 +0000 UTC m=+921.039177209" lastFinishedPulling="2026-02-03 09:25:46.54216528 +0000 UTC m=+937.692632655" observedRunningTime="2026-02-03 09:25:58.513133465 +0000 UTC m=+949.663600840" watchObservedRunningTime="2026-02-03 09:25:58.518957587 +0000 UTC m=+949.669424962" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.542009 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" podStartSLOduration=12.853244387 podStartE2EDuration="30.541987757s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:29.555858935 +0000 UTC m=+920.706326310" lastFinishedPulling="2026-02-03 09:25:47.244602295 +0000 UTC m=+938.395069680" observedRunningTime="2026-02-03 09:25:58.541699548 +0000 UTC m=+949.692166923" watchObservedRunningTime="2026-02-03 09:25:58.541987757 +0000 UTC m=+949.692455132" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.568391 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" podStartSLOduration=14.459306022 podStartE2EDuration="30.568374981s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.429742496 +0000 UTC m=+921.580209861" lastFinishedPulling="2026-02-03 09:25:46.538811445 +0000 UTC m=+937.689278820" observedRunningTime="2026-02-03 09:25:58.565741898 +0000 UTC m=+949.716209273" watchObservedRunningTime="2026-02-03 09:25:58.568374981 +0000 UTC m=+949.718842356" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.589598 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" podStartSLOduration=13.730282546 podStartE2EDuration="30.589574783s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.384976898 +0000 UTC m=+921.535444273" lastFinishedPulling="2026-02-03 09:25:47.244269135 +0000 UTC m=+938.394736510" observedRunningTime="2026-02-03 09:25:58.58212374 +0000 UTC m=+949.732591115" watchObservedRunningTime="2026-02-03 09:25:58.589574783 +0000 UTC m=+949.740042168" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.665671 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" podStartSLOduration=14.515569721 podStartE2EDuration="30.665658151s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.400757451 +0000 UTC m=+921.551224826" lastFinishedPulling="2026-02-03 09:25:46.550845881 +0000 UTC m=+937.701313256" observedRunningTime="2026-02-03 09:25:58.647218484 +0000 UTC m=+949.797685859" watchObservedRunningTime="2026-02-03 09:25:58.665658151 +0000 UTC m=+949.816125516" Feb 03 09:25:58 crc kubenswrapper[4756]: I0203 09:25:58.723870 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lqghx" podStartSLOduration=3.14584734 podStartE2EDuration="29.723849968s" podCreationTimestamp="2026-02-03 09:25:29 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.449898406 +0000 UTC m=+921.600365781" lastFinishedPulling="2026-02-03 09:25:57.027901034 +0000 UTC m=+948.178368409" observedRunningTime="2026-02-03 09:25:58.72038015 +0000 UTC m=+949.870847525" watchObservedRunningTime="2026-02-03 09:25:58.723849968 +0000 UTC m=+949.874317343" Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.414394 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" event={"ID":"999c5dd9-6dbb-44b9-bdd7-52c146d34e87","Type":"ContainerStarted","Data":"e2fa1264347c60cc7c5fc89f52055e2e1ae06a08929970ecf1a48102f3dcdbca"} Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.415595 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.417274 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" event={"ID":"919fda51-9a13-4233-ba62-a85ac3559a02","Type":"ContainerStarted","Data":"38644b7b80b3f78034d358d0c33822ab0f39ab5538ec075b57a58d950a019a61"} Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.417846 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.419727 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" event={"ID":"d9f68e89-0c29-4d3d-9dae-7448ab35206f","Type":"ContainerStarted","Data":"a035b40e3ff71a19a1d534c7637d0f3f2e4384d54b3d407d5c3271b3ebc36375"} Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.420211 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.422232 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" event={"ID":"6dd1801b-ac57-4d28-9554-ef8ad7148c08","Type":"ContainerStarted","Data":"7cb19697478c7e4b5516a24b745050fd502c2fa56877e646525399423dab882a"} Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.423142 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.436543 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" podStartSLOduration=3.6452737109999998 podStartE2EDuration="31.436521363s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.416386949 +0000 UTC m=+921.566854324" lastFinishedPulling="2026-02-03 09:25:58.207634591 +0000 UTC m=+949.358101976" observedRunningTime="2026-02-03 09:25:59.430314939 +0000 UTC m=+950.580782314" watchObservedRunningTime="2026-02-03 09:25:59.436521363 +0000 UTC m=+950.586988728" Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.447823 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" podStartSLOduration=3.245875645 podStartE2EDuration="31.447801846s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:29.862475194 +0000 UTC m=+921.012942569" lastFinishedPulling="2026-02-03 09:25:58.064401395 +0000 UTC m=+949.214868770" observedRunningTime="2026-02-03 09:25:59.442377166 +0000 UTC m=+950.592844541" watchObservedRunningTime="2026-02-03 09:25:59.447801846 +0000 UTC m=+950.598269221" Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.464316 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" podStartSLOduration=3.255064691 podStartE2EDuration="31.464299651s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:29.854771284 +0000 UTC m=+921.005238659" lastFinishedPulling="2026-02-03 09:25:58.064006244 +0000 UTC m=+949.214473619" observedRunningTime="2026-02-03 09:25:59.458635814 +0000 UTC m=+950.609103189" watchObservedRunningTime="2026-02-03 09:25:59.464299651 +0000 UTC m=+950.614767026" Feb 03 09:25:59 crc kubenswrapper[4756]: I0203 09:25:59.477883 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" podStartSLOduration=3.7919917659999998 podStartE2EDuration="31.477865406s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:25:30.377803214 +0000 UTC m=+921.528270589" lastFinishedPulling="2026-02-03 09:25:58.063676854 +0000 UTC m=+949.214144229" observedRunningTime="2026-02-03 09:25:59.47416887 +0000 UTC m=+950.624636245" watchObservedRunningTime="2026-02-03 09:25:59.477865406 +0000 UTC m=+950.628332781" Feb 03 09:26:00 crc kubenswrapper[4756]: I0203 09:26:00.311538 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:26:00 crc kubenswrapper[4756]: I0203 09:26:00.318315 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aae19bbd-e8ec-4a6f-8e77-9dd46373e31b-cert\") pod \"infra-operator-controller-manager-79955696d6-vvkmq\" (UID: \"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:26:00 crc kubenswrapper[4756]: I0203 09:26:00.490143 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:26:00 crc kubenswrapper[4756]: I0203 09:26:00.717242 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:26:00 crc kubenswrapper[4756]: I0203 09:26:00.730392 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0ebd520-4c15-4c12-89fe-2eb8a407ca25-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx\" (UID: \"a0ebd520-4c15-4c12-89fe-2eb8a407ca25\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:26:00 crc kubenswrapper[4756]: I0203 09:26:00.807117 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:26:00 crc kubenswrapper[4756]: I0203 09:26:00.918768 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq"] Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.091282 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx"] Feb 03 09:26:01 crc kubenswrapper[4756]: W0203 09:26:01.097696 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0ebd520_4c15_4c12_89fe_2eb8a407ca25.slice/crio-fe9e235a6821037757b28cdb3b4983ec1a72127f9f24ba93caf3540ae5f7bf61 WatchSource:0}: Error finding container fe9e235a6821037757b28cdb3b4983ec1a72127f9f24ba93caf3540ae5f7bf61: Status 404 returned error can't find the container with id fe9e235a6821037757b28cdb3b4983ec1a72127f9f24ba93caf3540ae5f7bf61 Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.122822 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.122960 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.136445 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-metrics-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.136469 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2183ff26-c37d-472e-bf53-e0beba79a31d-webhook-certs\") pod \"openstack-operator-controller-manager-6c6ffddb96-rvk7w\" (UID: \"2183ff26-c37d-472e-bf53-e0beba79a31d\") " pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.289750 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.446916 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" event={"ID":"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b","Type":"ContainerStarted","Data":"0100aded072ee70d3ee9cb2c1ae3ad2e0246e37ca44f9d31e077d50605bcf098"} Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.448148 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" event={"ID":"a0ebd520-4c15-4c12-89fe-2eb8a407ca25","Type":"ContainerStarted","Data":"fe9e235a6821037757b28cdb3b4983ec1a72127f9f24ba93caf3540ae5f7bf61"} Feb 03 09:26:01 crc kubenswrapper[4756]: I0203 09:26:01.716396 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w"] Feb 03 09:26:01 crc kubenswrapper[4756]: W0203 09:26:01.730796 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2183ff26_c37d_472e_bf53_e0beba79a31d.slice/crio-90f4f8a0bd01f7d918092c4f02a2a7a80150aec4508f3dba93c408e814ba1d61 WatchSource:0}: Error finding container 90f4f8a0bd01f7d918092c4f02a2a7a80150aec4508f3dba93c408e814ba1d61: Status 404 returned error can't find the container with id 90f4f8a0bd01f7d918092c4f02a2a7a80150aec4508f3dba93c408e814ba1d61 Feb 03 09:26:02 crc kubenswrapper[4756]: I0203 09:26:02.457763 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" event={"ID":"2183ff26-c37d-472e-bf53-e0beba79a31d","Type":"ContainerStarted","Data":"c9fa47e75c28080306b19703efe3761f6de0ceea639df93b341cf5f96a4e8da5"} Feb 03 09:26:02 crc kubenswrapper[4756]: I0203 09:26:02.458130 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:26:02 crc kubenswrapper[4756]: I0203 09:26:02.458145 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" event={"ID":"2183ff26-c37d-472e-bf53-e0beba79a31d","Type":"ContainerStarted","Data":"90f4f8a0bd01f7d918092c4f02a2a7a80150aec4508f3dba93c408e814ba1d61"} Feb 03 09:26:02 crc kubenswrapper[4756]: I0203 09:26:02.483553 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" podStartSLOduration=34.483534028 podStartE2EDuration="34.483534028s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:26:02.481426332 +0000 UTC m=+953.631893707" watchObservedRunningTime="2026-02-03 09:26:02.483534028 +0000 UTC m=+953.634001403" Feb 03 09:26:04 crc kubenswrapper[4756]: I0203 09:26:04.469672 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" event={"ID":"a0ebd520-4c15-4c12-89fe-2eb8a407ca25","Type":"ContainerStarted","Data":"abe74f3c4a9a998a6b5655426e4c3c47c4eee8f6f627f6edd2e8403ab3fca466"} Feb 03 09:26:04 crc kubenswrapper[4756]: I0203 09:26:04.470091 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:26:04 crc kubenswrapper[4756]: I0203 09:26:04.471992 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" event={"ID":"aae19bbd-e8ec-4a6f-8e77-9dd46373e31b","Type":"ContainerStarted","Data":"ff464dc4058400a306f309b0afebe393f97a5f7ab5baf4c1b644c75d15654f78"} Feb 03 09:26:04 crc kubenswrapper[4756]: I0203 09:26:04.472369 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:26:04 crc kubenswrapper[4756]: I0203 09:26:04.502736 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" podStartSLOduration=33.940546824 podStartE2EDuration="36.502718311s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:26:01.099325502 +0000 UTC m=+952.249792877" lastFinishedPulling="2026-02-03 09:26:03.661496989 +0000 UTC m=+954.811964364" observedRunningTime="2026-02-03 09:26:04.497094406 +0000 UTC m=+955.647561781" watchObservedRunningTime="2026-02-03 09:26:04.502718311 +0000 UTC m=+955.653185686" Feb 03 09:26:04 crc kubenswrapper[4756]: I0203 09:26:04.513159 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" podStartSLOduration=33.783503447 podStartE2EDuration="36.513138716s" podCreationTimestamp="2026-02-03 09:25:28 +0000 UTC" firstStartedPulling="2026-02-03 09:26:00.936236738 +0000 UTC m=+952.086704113" lastFinishedPulling="2026-02-03 09:26:03.665872007 +0000 UTC m=+954.816339382" observedRunningTime="2026-02-03 09:26:04.511417103 +0000 UTC m=+955.661884498" watchObservedRunningTime="2026-02-03 09:26:04.513138716 +0000 UTC m=+955.663606091" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.539347 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-6pbkt" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.551497 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-dplk2" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.562656 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-qnklz" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.595454 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-pgdzr" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.616379 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-564mj" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.658785 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-hvnx5" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.701266 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-zhlxl" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.809538 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-w7vdf" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.816339 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-jwlpw" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.831105 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-5q5s7" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.929205 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-qnwj2" Feb 03 09:26:08 crc kubenswrapper[4756]: I0203 09:26:08.950066 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-bqthj" Feb 03 09:26:09 crc kubenswrapper[4756]: I0203 09:26:09.046516 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-wl9ft" Feb 03 09:26:09 crc kubenswrapper[4756]: I0203 09:26:09.111874 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-w87dt" Feb 03 09:26:09 crc kubenswrapper[4756]: I0203 09:26:09.143973 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-gn2d2" Feb 03 09:26:09 crc kubenswrapper[4756]: I0203 09:26:09.161859 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-2cw2v" Feb 03 09:26:09 crc kubenswrapper[4756]: I0203 09:26:09.258857 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-l5wms" Feb 03 09:26:09 crc kubenswrapper[4756]: I0203 09:26:09.335605 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-564965969-9rdzw" Feb 03 09:26:09 crc kubenswrapper[4756]: I0203 09:26:09.367418 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-7rrf6" Feb 03 09:26:10 crc kubenswrapper[4756]: I0203 09:26:10.496350 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79955696d6-vvkmq" Feb 03 09:26:10 crc kubenswrapper[4756]: I0203 09:26:10.814213 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.254157 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kdwbl"] Feb 03 09:26:11 crc kubenswrapper[4756]: E0203 09:26:11.256366 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerName="extract-content" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.256388 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerName="extract-content" Feb 03 09:26:11 crc kubenswrapper[4756]: E0203 09:26:11.256400 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerName="registry-server" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.256408 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerName="registry-server" Feb 03 09:26:11 crc kubenswrapper[4756]: E0203 09:26:11.256457 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerName="extract-utilities" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.256467 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerName="extract-utilities" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.256659 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="64eb5569-95ab-4135-9d4f-4bda9904b174" containerName="registry-server" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.260858 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.266870 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kdwbl"] Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.309901 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6c6ffddb96-rvk7w" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.366193 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2qlm\" (UniqueName: \"kubernetes.io/projected/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-kube-api-access-p2qlm\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.366529 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-utilities\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.366560 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-catalog-content\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.467941 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-utilities\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.467986 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-catalog-content\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.468034 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2qlm\" (UniqueName: \"kubernetes.io/projected/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-kube-api-access-p2qlm\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.468490 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-utilities\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.468603 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-catalog-content\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.493858 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2qlm\" (UniqueName: \"kubernetes.io/projected/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-kube-api-access-p2qlm\") pod \"community-operators-kdwbl\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:11 crc kubenswrapper[4756]: I0203 09:26:11.584071 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:12 crc kubenswrapper[4756]: I0203 09:26:12.173137 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kdwbl"] Feb 03 09:26:12 crc kubenswrapper[4756]: I0203 09:26:12.520731 4756 generic.go:334] "Generic (PLEG): container finished" podID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerID="b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471" exitCode=0 Feb 03 09:26:12 crc kubenswrapper[4756]: I0203 09:26:12.520800 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdwbl" event={"ID":"6e6a34df-047f-4137-bf61-5ba9ecc96bc0","Type":"ContainerDied","Data":"b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471"} Feb 03 09:26:12 crc kubenswrapper[4756]: I0203 09:26:12.521008 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdwbl" event={"ID":"6e6a34df-047f-4137-bf61-5ba9ecc96bc0","Type":"ContainerStarted","Data":"1afe4d9ad67c31b716dc06200db2d7321b86e3feefc10114db63e3834b6ef842"} Feb 03 09:26:13 crc kubenswrapper[4756]: I0203 09:26:13.529231 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdwbl" event={"ID":"6e6a34df-047f-4137-bf61-5ba9ecc96bc0","Type":"ContainerStarted","Data":"671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602"} Feb 03 09:26:14 crc kubenswrapper[4756]: I0203 09:26:14.537560 4756 generic.go:334] "Generic (PLEG): container finished" podID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerID="671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602" exitCode=0 Feb 03 09:26:14 crc kubenswrapper[4756]: I0203 09:26:14.537697 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdwbl" event={"ID":"6e6a34df-047f-4137-bf61-5ba9ecc96bc0","Type":"ContainerDied","Data":"671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602"} Feb 03 09:26:15 crc kubenswrapper[4756]: I0203 09:26:15.547210 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdwbl" event={"ID":"6e6a34df-047f-4137-bf61-5ba9ecc96bc0","Type":"ContainerStarted","Data":"7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8"} Feb 03 09:26:15 crc kubenswrapper[4756]: I0203 09:26:15.566275 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kdwbl" podStartSLOduration=1.881293004 podStartE2EDuration="4.566257297s" podCreationTimestamp="2026-02-03 09:26:11 +0000 UTC" firstStartedPulling="2026-02-03 09:26:12.522290897 +0000 UTC m=+963.672758272" lastFinishedPulling="2026-02-03 09:26:15.20725518 +0000 UTC m=+966.357722565" observedRunningTime="2026-02-03 09:26:15.560523648 +0000 UTC m=+966.710991043" watchObservedRunningTime="2026-02-03 09:26:15.566257297 +0000 UTC m=+966.716724672" Feb 03 09:26:21 crc kubenswrapper[4756]: I0203 09:26:21.585035 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:21 crc kubenswrapper[4756]: I0203 09:26:21.585607 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:21 crc kubenswrapper[4756]: I0203 09:26:21.637413 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:22 crc kubenswrapper[4756]: I0203 09:26:22.627431 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:22 crc kubenswrapper[4756]: I0203 09:26:22.670523 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kdwbl"] Feb 03 09:26:24 crc kubenswrapper[4756]: I0203 09:26:24.601867 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kdwbl" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerName="registry-server" containerID="cri-o://7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8" gracePeriod=2 Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.453194 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.545345 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2qlm\" (UniqueName: \"kubernetes.io/projected/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-kube-api-access-p2qlm\") pod \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.545564 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-catalog-content\") pod \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.545599 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-utilities\") pod \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\" (UID: \"6e6a34df-047f-4137-bf61-5ba9ecc96bc0\") " Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.546536 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-utilities" (OuterVolumeSpecName: "utilities") pod "6e6a34df-047f-4137-bf61-5ba9ecc96bc0" (UID: "6e6a34df-047f-4137-bf61-5ba9ecc96bc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.550906 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-kube-api-access-p2qlm" (OuterVolumeSpecName: "kube-api-access-p2qlm") pod "6e6a34df-047f-4137-bf61-5ba9ecc96bc0" (UID: "6e6a34df-047f-4137-bf61-5ba9ecc96bc0"). InnerVolumeSpecName "kube-api-access-p2qlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.593963 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e6a34df-047f-4137-bf61-5ba9ecc96bc0" (UID: "6e6a34df-047f-4137-bf61-5ba9ecc96bc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.608670 4756 generic.go:334] "Generic (PLEG): container finished" podID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerID="7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8" exitCode=0 Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.608724 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdwbl" event={"ID":"6e6a34df-047f-4137-bf61-5ba9ecc96bc0","Type":"ContainerDied","Data":"7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8"} Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.608771 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdwbl" event={"ID":"6e6a34df-047f-4137-bf61-5ba9ecc96bc0","Type":"ContainerDied","Data":"1afe4d9ad67c31b716dc06200db2d7321b86e3feefc10114db63e3834b6ef842"} Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.608795 4756 scope.go:117] "RemoveContainer" containerID="7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.608927 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdwbl" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.635808 4756 scope.go:117] "RemoveContainer" containerID="671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.638508 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kdwbl"] Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.645841 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kdwbl"] Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.646608 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2qlm\" (UniqueName: \"kubernetes.io/projected/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-kube-api-access-p2qlm\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.646639 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.646650 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e6a34df-047f-4137-bf61-5ba9ecc96bc0-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.655596 4756 scope.go:117] "RemoveContainer" containerID="b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.679700 4756 scope.go:117] "RemoveContainer" containerID="7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8" Feb 03 09:26:25 crc kubenswrapper[4756]: E0203 09:26:25.680156 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8\": container with ID starting with 7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8 not found: ID does not exist" containerID="7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.680214 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8"} err="failed to get container status \"7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8\": rpc error: code = NotFound desc = could not find container \"7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8\": container with ID starting with 7598c535ea0119314916f5e3353c7391d2a2f99883673fb8ed587f2d65a5d6f8 not found: ID does not exist" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.680242 4756 scope.go:117] "RemoveContainer" containerID="671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602" Feb 03 09:26:25 crc kubenswrapper[4756]: E0203 09:26:25.680641 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602\": container with ID starting with 671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602 not found: ID does not exist" containerID="671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.680666 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602"} err="failed to get container status \"671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602\": rpc error: code = NotFound desc = could not find container \"671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602\": container with ID starting with 671a4b0547c86773039fe3bd2983fd37d8ac51a76c744b626290bb99eec00602 not found: ID does not exist" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.680678 4756 scope.go:117] "RemoveContainer" containerID="b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471" Feb 03 09:26:25 crc kubenswrapper[4756]: E0203 09:26:25.681039 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471\": container with ID starting with b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471 not found: ID does not exist" containerID="b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471" Feb 03 09:26:25 crc kubenswrapper[4756]: I0203 09:26:25.681088 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471"} err="failed to get container status \"b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471\": rpc error: code = NotFound desc = could not find container \"b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471\": container with ID starting with b16bde55f5586cbd70322853bb147cfb70bbb4aca817ccb592bc6b0cdd26d471 not found: ID does not exist" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.296362 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kkwtq"] Feb 03 09:26:26 crc kubenswrapper[4756]: E0203 09:26:26.296692 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerName="extract-content" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.296710 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerName="extract-content" Feb 03 09:26:26 crc kubenswrapper[4756]: E0203 09:26:26.296721 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerName="registry-server" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.296728 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerName="registry-server" Feb 03 09:26:26 crc kubenswrapper[4756]: E0203 09:26:26.296737 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerName="extract-utilities" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.296743 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerName="extract-utilities" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.296881 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" containerName="registry-server" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.300045 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.303682 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.304340 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.304518 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-xxcpc" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.304690 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.306155 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kkwtq"] Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.352867 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2rpqm"] Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.353963 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.359329 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.360133 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-config\") pod \"dnsmasq-dns-675f4bcbfc-kkwtq\" (UID: \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.360251 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvfbl\" (UniqueName: \"kubernetes.io/projected/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-kube-api-access-lvfbl\") pod \"dnsmasq-dns-675f4bcbfc-kkwtq\" (UID: \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.369515 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2rpqm"] Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.461237 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.461314 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-config\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.461350 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvfbl\" (UniqueName: \"kubernetes.io/projected/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-kube-api-access-lvfbl\") pod \"dnsmasq-dns-675f4bcbfc-kkwtq\" (UID: \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.461394 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52vlp\" (UniqueName: \"kubernetes.io/projected/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-kube-api-access-52vlp\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.461436 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-config\") pod \"dnsmasq-dns-675f4bcbfc-kkwtq\" (UID: \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.462328 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-config\") pod \"dnsmasq-dns-675f4bcbfc-kkwtq\" (UID: \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.479308 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvfbl\" (UniqueName: \"kubernetes.io/projected/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-kube-api-access-lvfbl\") pod \"dnsmasq-dns-675f4bcbfc-kkwtq\" (UID: \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.563026 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.563105 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-config\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.563159 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52vlp\" (UniqueName: \"kubernetes.io/projected/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-kube-api-access-52vlp\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.563992 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.564016 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-config\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.577279 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52vlp\" (UniqueName: \"kubernetes.io/projected/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-kube-api-access-52vlp\") pod \"dnsmasq-dns-78dd6ddcc-2rpqm\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.617874 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.673886 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:26 crc kubenswrapper[4756]: I0203 09:26:26.922562 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2rpqm"] Feb 03 09:26:26 crc kubenswrapper[4756]: W0203 09:26:26.931104 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bcab1be_ec76_463a_b5e7_f5c8e0f78338.slice/crio-84257d69a73a0db3804e395846d1dc33cb88d5935a793a6de219716cd8f7afb4 WatchSource:0}: Error finding container 84257d69a73a0db3804e395846d1dc33cb88d5935a793a6de219716cd8f7afb4: Status 404 returned error can't find the container with id 84257d69a73a0db3804e395846d1dc33cb88d5935a793a6de219716cd8f7afb4 Feb 03 09:26:27 crc kubenswrapper[4756]: I0203 09:26:27.045649 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kkwtq"] Feb 03 09:26:27 crc kubenswrapper[4756]: I0203 09:26:27.624094 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e6a34df-047f-4137-bf61-5ba9ecc96bc0" path="/var/lib/kubelet/pods/6e6a34df-047f-4137-bf61-5ba9ecc96bc0/volumes" Feb 03 09:26:27 crc kubenswrapper[4756]: I0203 09:26:27.624902 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" event={"ID":"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae","Type":"ContainerStarted","Data":"e0861a08363c2ccf8571f62eaa73adf076c097138ee6774b6787b13bf4060fa4"} Feb 03 09:26:27 crc kubenswrapper[4756]: I0203 09:26:27.625760 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" event={"ID":"3bcab1be-ec76-463a-b5e7-f5c8e0f78338","Type":"ContainerStarted","Data":"84257d69a73a0db3804e395846d1dc33cb88d5935a793a6de219716cd8f7afb4"} Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.120971 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kkwtq"] Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.148700 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9sv9k"] Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.150034 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.174360 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9sv9k"] Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.208018 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.208109 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-config\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.208201 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkttw\" (UniqueName: \"kubernetes.io/projected/84bfa317-7260-46d2-8e43-13fffac7d17f-kube-api-access-nkttw\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.309851 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkttw\" (UniqueName: \"kubernetes.io/projected/84bfa317-7260-46d2-8e43-13fffac7d17f-kube-api-access-nkttw\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.309930 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.309981 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-config\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.312039 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-config\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.312214 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.335494 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkttw\" (UniqueName: \"kubernetes.io/projected/84bfa317-7260-46d2-8e43-13fffac7d17f-kube-api-access-nkttw\") pod \"dnsmasq-dns-666b6646f7-9sv9k\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.478111 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.486996 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2rpqm"] Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.514257 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwr6s"] Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.516498 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.563042 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwr6s"] Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.621364 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqpwk\" (UniqueName: \"kubernetes.io/projected/d1dbab7f-1fdd-49f6-a7a6-904feb481959-kube-api-access-cqpwk\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.624647 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-config\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.625960 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.727282 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.727519 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqpwk\" (UniqueName: \"kubernetes.io/projected/d1dbab7f-1fdd-49f6-a7a6-904feb481959-kube-api-access-cqpwk\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.727573 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-config\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.728270 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-config\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.728937 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.754844 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqpwk\" (UniqueName: \"kubernetes.io/projected/d1dbab7f-1fdd-49f6-a7a6-904feb481959-kube-api-access-cqpwk\") pod \"dnsmasq-dns-57d769cc4f-vwr6s\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:29 crc kubenswrapper[4756]: I0203 09:26:29.900131 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.055953 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9sv9k"] Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.336828 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.338090 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.341012 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.341301 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.341580 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.341814 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.342082 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.342495 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.342562 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hr92v" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.374976 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.425508 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwr6s"] Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.436864 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.436924 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.436960 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.436986 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.437026 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlwzr\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-kube-api-access-xlwzr\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.437052 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.437073 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-config-data\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.437160 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.437190 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.437223 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.437248 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538545 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538614 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538634 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538652 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538742 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538759 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538808 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538830 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538876 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlwzr\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-kube-api-access-xlwzr\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538895 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.538910 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-config-data\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.543438 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-config-data\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.543480 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.547428 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.547788 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.548142 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.548224 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.551657 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.552230 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.557402 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.558496 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.558867 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlwzr\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-kube-api-access-xlwzr\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.575920 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.653848 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.662018 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" event={"ID":"84bfa317-7260-46d2-8e43-13fffac7d17f","Type":"ContainerStarted","Data":"3a051ad86d9331b6ac540b142b0881c6eac993427437218be93e3bd1ea2de735"} Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.662154 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.665805 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.666242 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.666386 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.666543 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.666644 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.666739 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.667189 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qzgw2" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.680692 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.686815 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742496 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742567 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742595 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742619 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742656 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742682 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742717 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742745 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742783 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742813 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.742850 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72vpb\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-kube-api-access-72vpb\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844164 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844205 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844242 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844261 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844292 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844315 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844355 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72vpb\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-kube-api-access-72vpb\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844380 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844405 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844425 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844443 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.844567 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.845123 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.845589 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.845714 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.845922 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.846431 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.848939 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.850214 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.859665 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.866257 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72vpb\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-kube-api-access-72vpb\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.869198 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:30 crc kubenswrapper[4756]: I0203 09:26:30.906238 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:31 crc kubenswrapper[4756]: I0203 09:26:31.018546 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:26:31 crc kubenswrapper[4756]: I0203 09:26:31.995932 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.001030 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.005989 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.006203 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.006351 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-n62sh" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.009528 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.014976 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.018430 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.062291 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-kolla-config\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.062356 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-config-data-default\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.062386 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.062421 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.062460 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfj8g\" (UniqueName: \"kubernetes.io/projected/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-kube-api-access-bfj8g\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.062482 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.062505 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.062522 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.163556 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-kolla-config\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.163631 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-config-data-default\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.163670 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.163711 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.163733 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfj8g\" (UniqueName: \"kubernetes.io/projected/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-kube-api-access-bfj8g\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.163761 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.163791 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.163812 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.164211 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-kolla-config\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.164422 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.165081 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-config-data-default\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.165156 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.165248 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.171047 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.171799 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.185852 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfj8g\" (UniqueName: \"kubernetes.io/projected/77baa89d-5260-4b1d-a633-a4f3e37ed6bb-kube-api-access-bfj8g\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.201553 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"77baa89d-5260-4b1d-a633-a4f3e37ed6bb\") " pod="openstack/openstack-galera-0" Feb 03 09:26:32 crc kubenswrapper[4756]: I0203 09:26:32.324960 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.399612 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.404126 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.408570 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.408613 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.408685 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-gf9hl" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.408802 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.410348 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.449962 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjjw"] Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.453422 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.462173 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjjw"] Feb 03 09:26:33 crc kubenswrapper[4756]: W0203 09:26:33.553990 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1dbab7f_1fdd_49f6_a7a6_904feb481959.slice/crio-8d34da076b63a07b58e3b4aa772da20d4c1dc54d4f25db8d8d5e38dc8b2beeee WatchSource:0}: Error finding container 8d34da076b63a07b58e3b4aa772da20d4c1dc54d4f25db8d8d5e38dc8b2beeee: Status 404 returned error can't find the container with id 8d34da076b63a07b58e3b4aa772da20d4c1dc54d4f25db8d8d5e38dc8b2beeee Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.582399 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.582467 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-catalog-content\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.582503 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-utilities\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.582526 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.582859 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mml57\" (UniqueName: \"kubernetes.io/projected/8d6139e2-68a5-4097-b8c2-cf11d11ed241-kube-api-access-mml57\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.582927 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.582969 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.583037 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.583088 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.583133 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.583170 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdjcq\" (UniqueName: \"kubernetes.io/projected/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-kube-api-access-qdjcq\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.684215 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" event={"ID":"d1dbab7f-1fdd-49f6-a7a6-904feb481959","Type":"ContainerStarted","Data":"8d34da076b63a07b58e3b4aa772da20d4c1dc54d4f25db8d8d5e38dc8b2beeee"} Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685573 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685624 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685651 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685676 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685697 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdjcq\" (UniqueName: \"kubernetes.io/projected/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-kube-api-access-qdjcq\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685739 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685762 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-catalog-content\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685792 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-utilities\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685812 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685833 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mml57\" (UniqueName: \"kubernetes.io/projected/8d6139e2-68a5-4097-b8c2-cf11d11ed241-kube-api-access-mml57\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.685855 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.686282 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.686527 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.686532 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.686578 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-catalog-content\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.687147 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-utilities\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.687371 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.688058 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.695085 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.706183 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.711630 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mml57\" (UniqueName: \"kubernetes.io/projected/8d6139e2-68a5-4097-b8c2-cf11d11ed241-kube-api-access-mml57\") pod \"redhat-marketplace-qcjjw\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.715993 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdjcq\" (UniqueName: \"kubernetes.io/projected/47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c-kube-api-access-qdjcq\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.720649 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c\") " pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.735745 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.781798 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.792014 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.793122 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.815388 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.815724 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vwsqh" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.815925 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.861592 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.915622 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.915684 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-config-data\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.915737 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-kolla-config\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.915916 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:33 crc kubenswrapper[4756]: I0203 09:26:33.916067 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9fc8\" (UniqueName: \"kubernetes.io/projected/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-kube-api-access-q9fc8\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.017244 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9fc8\" (UniqueName: \"kubernetes.io/projected/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-kube-api-access-q9fc8\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.017327 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.017345 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-config-data\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.017385 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-kolla-config\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.017417 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.018693 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-kolla-config\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.019919 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-config-data\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.022285 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.036314 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.046402 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9fc8\" (UniqueName: \"kubernetes.io/projected/ce92a835-f667-4f5b-b9e1-c8afcf476b9a-kube-api-access-q9fc8\") pod \"memcached-0\" (UID: \"ce92a835-f667-4f5b-b9e1-c8afcf476b9a\") " pod="openstack/memcached-0" Feb 03 09:26:34 crc kubenswrapper[4756]: I0203 09:26:34.137803 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 03 09:26:35 crc kubenswrapper[4756]: I0203 09:26:35.778850 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:26:35 crc kubenswrapper[4756]: I0203 09:26:35.780058 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 03 09:26:35 crc kubenswrapper[4756]: I0203 09:26:35.782718 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-lj2dc" Feb 03 09:26:35 crc kubenswrapper[4756]: I0203 09:26:35.798174 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:26:35 crc kubenswrapper[4756]: I0203 09:26:35.951786 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbwqv\" (UniqueName: \"kubernetes.io/projected/773ac9a0-544a-4c28-a100-59f10ec789c7-kube-api-access-qbwqv\") pod \"kube-state-metrics-0\" (UID: \"773ac9a0-544a-4c28-a100-59f10ec789c7\") " pod="openstack/kube-state-metrics-0" Feb 03 09:26:36 crc kubenswrapper[4756]: I0203 09:26:36.053793 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbwqv\" (UniqueName: \"kubernetes.io/projected/773ac9a0-544a-4c28-a100-59f10ec789c7-kube-api-access-qbwqv\") pod \"kube-state-metrics-0\" (UID: \"773ac9a0-544a-4c28-a100-59f10ec789c7\") " pod="openstack/kube-state-metrics-0" Feb 03 09:26:36 crc kubenswrapper[4756]: I0203 09:26:36.091493 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbwqv\" (UniqueName: \"kubernetes.io/projected/773ac9a0-544a-4c28-a100-59f10ec789c7-kube-api-access-qbwqv\") pod \"kube-state-metrics-0\" (UID: \"773ac9a0-544a-4c28-a100-59f10ec789c7\") " pod="openstack/kube-state-metrics-0" Feb 03 09:26:36 crc kubenswrapper[4756]: I0203 09:26:36.107353 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.344800 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-wnf95"] Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.346266 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.348718 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-szcnl" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.349029 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.352639 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.359251 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wnf95"] Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.397114 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-fb566"] Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.399292 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.405411 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fb566"] Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.513854 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-run\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.513908 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15695b6e-11fa-4913-b502-a78974ddf81c-scripts\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.513923 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/15695b6e-11fa-4913-b502-a78974ddf81c-ovn-controller-tls-certs\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514064 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-scripts\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514142 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-run\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514500 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-lib\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514535 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gx7w\" (UniqueName: \"kubernetes.io/projected/15695b6e-11fa-4913-b502-a78974ddf81c-kube-api-access-7gx7w\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514643 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-log-ovn\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514706 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-etc-ovs\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514797 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfxdc\" (UniqueName: \"kubernetes.io/projected/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-kube-api-access-wfxdc\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514818 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-run-ovn\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514832 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15695b6e-11fa-4913-b502-a78974ddf81c-combined-ca-bundle\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.514937 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-log\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.615717 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-log\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.615776 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-run\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.615809 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15695b6e-11fa-4913-b502-a78974ddf81c-scripts\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.615833 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/15695b6e-11fa-4913-b502-a78974ddf81c-ovn-controller-tls-certs\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.615882 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-scripts\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.615925 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-run\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.615957 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-lib\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.615977 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gx7w\" (UniqueName: \"kubernetes.io/projected/15695b6e-11fa-4913-b502-a78974ddf81c-kube-api-access-7gx7w\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616002 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-log-ovn\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616025 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-etc-ovs\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616067 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfxdc\" (UniqueName: \"kubernetes.io/projected/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-kube-api-access-wfxdc\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616090 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-run-ovn\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616110 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15695b6e-11fa-4913-b502-a78974ddf81c-combined-ca-bundle\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616291 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-log\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616462 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-run\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616529 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-var-lib\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.616631 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-etc-ovs\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.617071 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-log-ovn\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.618241 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-run\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.618478 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/15695b6e-11fa-4913-b502-a78974ddf81c-var-run-ovn\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.620231 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15695b6e-11fa-4913-b502-a78974ddf81c-scripts\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.620970 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-scripts\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.623361 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/15695b6e-11fa-4913-b502-a78974ddf81c-ovn-controller-tls-certs\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.623685 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15695b6e-11fa-4913-b502-a78974ddf81c-combined-ca-bundle\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.633719 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gx7w\" (UniqueName: \"kubernetes.io/projected/15695b6e-11fa-4913-b502-a78974ddf81c-kube-api-access-7gx7w\") pod \"ovn-controller-wnf95\" (UID: \"15695b6e-11fa-4913-b502-a78974ddf81c\") " pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.634233 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfxdc\" (UniqueName: \"kubernetes.io/projected/4600dd3a-bdd0-44b1-8a6d-f34313d5df5d-kube-api-access-wfxdc\") pod \"ovn-controller-ovs-fb566\" (UID: \"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d\") " pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.678120 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95" Feb 03 09:26:39 crc kubenswrapper[4756]: I0203 09:26:39.719727 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.196288 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.199064 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.203100 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.203265 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-774nv" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.203350 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.203403 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.204702 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.211066 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.327106 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z4jl\" (UniqueName: \"kubernetes.io/projected/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-kube-api-access-8z4jl\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.327644 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.327756 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.327828 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.327928 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-config\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.328004 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.328107 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.328140 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.429838 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z4jl\" (UniqueName: \"kubernetes.io/projected/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-kube-api-access-8z4jl\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.429978 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.430018 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.430047 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.430085 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-config\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.430127 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.430401 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.431316 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.432432 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.433085 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-config\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.433765 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.433810 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.436068 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.444065 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.444143 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.447680 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z4jl\" (UniqueName: \"kubernetes.io/projected/a2da6bae-fd2d-4f9c-b2aa-440d4264f17a-kube-api-access-8z4jl\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.458287 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a\") " pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:40 crc kubenswrapper[4756]: I0203 09:26:40.525834 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: E0203 09:26:42.676791 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 03 09:26:42 crc kubenswrapper[4756]: E0203 09:26:42.677382 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-52vlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-2rpqm_openstack(3bcab1be-ec76-463a-b5e7-f5c8e0f78338): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:26:42 crc kubenswrapper[4756]: E0203 09:26:42.678588 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" podUID="3bcab1be-ec76-463a-b5e7-f5c8e0f78338" Feb 03 09:26:42 crc kubenswrapper[4756]: E0203 09:26:42.679280 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 03 09:26:42 crc kubenswrapper[4756]: E0203 09:26:42.679415 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lvfbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-kkwtq_openstack(b58a001a-2f8b-4dfb-acb5-eecbdd2761ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:26:42 crc kubenswrapper[4756]: E0203 09:26:42.680670 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" podUID="b58a001a-2f8b-4dfb-acb5-eecbdd2761ae" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.772972 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.774251 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.776078 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.776251 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.776258 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-pnz42" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.776470 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.780561 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.974680 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.974767 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrf2m\" (UniqueName: \"kubernetes.io/projected/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-kube-api-access-vrf2m\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.974797 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.974814 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.974833 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.974862 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.974882 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:42 crc kubenswrapper[4756]: I0203 09:26:42.974909 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.076212 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrf2m\" (UniqueName: \"kubernetes.io/projected/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-kube-api-access-vrf2m\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.076495 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.076611 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.076720 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.076842 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.076957 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.077065 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.077196 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.077707 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.078952 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.081148 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.081711 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.096692 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.141653 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.141787 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.144711 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrf2m\" (UniqueName: \"kubernetes.io/projected/bfba7eb1-da32-4bc6-8309-f076ad7b32a0-kube-api-access-vrf2m\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.150171 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bfba7eb1-da32-4bc6-8309-f076ad7b32a0\") " pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.170278 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.356313 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.362878 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.490027 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-config\") pod \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\" (UID: \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\") " Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.490085 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52vlp\" (UniqueName: \"kubernetes.io/projected/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-kube-api-access-52vlp\") pod \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.490132 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvfbl\" (UniqueName: \"kubernetes.io/projected/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-kube-api-access-lvfbl\") pod \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\" (UID: \"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae\") " Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.490168 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-config\") pod \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.490329 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-dns-svc\") pod \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\" (UID: \"3bcab1be-ec76-463a-b5e7-f5c8e0f78338\") " Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.491258 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3bcab1be-ec76-463a-b5e7-f5c8e0f78338" (UID: "3bcab1be-ec76-463a-b5e7-f5c8e0f78338"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.492278 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-config" (OuterVolumeSpecName: "config") pod "b58a001a-2f8b-4dfb-acb5-eecbdd2761ae" (UID: "b58a001a-2f8b-4dfb-acb5-eecbdd2761ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.493601 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-config" (OuterVolumeSpecName: "config") pod "3bcab1be-ec76-463a-b5e7-f5c8e0f78338" (UID: "3bcab1be-ec76-463a-b5e7-f5c8e0f78338"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.496737 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-kube-api-access-52vlp" (OuterVolumeSpecName: "kube-api-access-52vlp") pod "3bcab1be-ec76-463a-b5e7-f5c8e0f78338" (UID: "3bcab1be-ec76-463a-b5e7-f5c8e0f78338"). InnerVolumeSpecName "kube-api-access-52vlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.496878 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-kube-api-access-lvfbl" (OuterVolumeSpecName: "kube-api-access-lvfbl") pod "b58a001a-2f8b-4dfb-acb5-eecbdd2761ae" (UID: "b58a001a-2f8b-4dfb-acb5-eecbdd2761ae"). InnerVolumeSpecName "kube-api-access-lvfbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.592174 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.592220 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.592235 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52vlp\" (UniqueName: \"kubernetes.io/projected/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-kube-api-access-52vlp\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.592250 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvfbl\" (UniqueName: \"kubernetes.io/projected/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae-kube-api-access-lvfbl\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.592262 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bcab1be-ec76-463a-b5e7-f5c8e0f78338-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.690704 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.700734 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.713249 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.720862 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.740576 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 09:26:43 crc kubenswrapper[4756]: W0203 09:26:43.744081 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod773ac9a0_544a_4c28_a100_59f10ec789c7.slice/crio-14a37fb11a14511a909627cd2e65b9d03bc79e6d4ea2e09bdf33e92e931cf813 WatchSource:0}: Error finding container 14a37fb11a14511a909627cd2e65b9d03bc79e6d4ea2e09bdf33e92e931cf813: Status 404 returned error can't find the container with id 14a37fb11a14511a909627cd2e65b9d03bc79e6d4ea2e09bdf33e92e931cf813 Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.768297 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.806171 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.815333 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ce92a835-f667-4f5b-b9e1-c8afcf476b9a","Type":"ContainerStarted","Data":"aa49c82681b4f641d47ab6e285db107371158157247d1ce160365b10b629bf85"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.821798 4756 generic.go:334] "Generic (PLEG): container finished" podID="84bfa317-7260-46d2-8e43-13fffac7d17f" containerID="4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0" exitCode=0 Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.821887 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" event={"ID":"84bfa317-7260-46d2-8e43-13fffac7d17f","Type":"ContainerDied","Data":"4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.829479 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" event={"ID":"b58a001a-2f8b-4dfb-acb5-eecbdd2761ae","Type":"ContainerDied","Data":"e0861a08363c2ccf8571f62eaa73adf076c097138ee6774b6787b13bf4060fa4"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.829661 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kkwtq" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.832372 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f2aeafe0-b129-44d5-86d9-0a95a7ac180d","Type":"ContainerStarted","Data":"1eca17f1e9bb4fb18827ebe59224016c1b128d8aa09d0a8610ed79914da98851"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.843272 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"773ac9a0-544a-4c28-a100-59f10ec789c7","Type":"ContainerStarted","Data":"14a37fb11a14511a909627cd2e65b9d03bc79e6d4ea2e09bdf33e92e931cf813"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.846827 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wnf95"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.859170 4756 generic.go:334] "Generic (PLEG): container finished" podID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" containerID="de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0" exitCode=0 Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.859494 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" event={"ID":"d1dbab7f-1fdd-49f6-a7a6-904feb481959","Type":"ContainerDied","Data":"de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.868296 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"77baa89d-5260-4b1d-a633-a4f3e37ed6bb","Type":"ContainerStarted","Data":"dfed340f9ef73dee0f74d33e30a570bec8ecac0d11c08cf9aebb083dd669f8f1"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.868950 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjjw"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.875418 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c","Type":"ContainerStarted","Data":"8feab9b6921feea3dc1c80e8e6eeb237b6abeb956563b87b52840f8dda6a8509"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.889520 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" event={"ID":"3bcab1be-ec76-463a-b5e7-f5c8e0f78338","Type":"ContainerDied","Data":"84257d69a73a0db3804e395846d1dc33cb88d5935a793a6de219716cd8f7afb4"} Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.889603 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2rpqm" Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.904461 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fb566"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.936312 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kkwtq"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.955789 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kkwtq"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.970505 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2rpqm"] Feb 03 09:26:43 crc kubenswrapper[4756]: I0203 09:26:43.976010 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2rpqm"] Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.011052 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 03 09:26:44 crc kubenswrapper[4756]: W0203 09:26:44.026075 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfba7eb1_da32_4bc6_8309_f076ad7b32a0.slice/crio-dc935433bafd4f7874e97531618336a524f13afe9a5bdc23387dd72162f1de5b WatchSource:0}: Error finding container dc935433bafd4f7874e97531618336a524f13afe9a5bdc23387dd72162f1de5b: Status 404 returned error can't find the container with id dc935433bafd4f7874e97531618336a524f13afe9a5bdc23387dd72162f1de5b Feb 03 09:26:44 crc kubenswrapper[4756]: E0203 09:26:44.162375 4756 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 03 09:26:44 crc kubenswrapper[4756]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/84bfa317-7260-46d2-8e43-13fffac7d17f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 03 09:26:44 crc kubenswrapper[4756]: > podSandboxID="3a051ad86d9331b6ac540b142b0881c6eac993427437218be93e3bd1ea2de735" Feb 03 09:26:44 crc kubenswrapper[4756]: E0203 09:26:44.162567 4756 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 03 09:26:44 crc kubenswrapper[4756]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nkttw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-9sv9k_openstack(84bfa317-7260-46d2-8e43-13fffac7d17f): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/84bfa317-7260-46d2-8e43-13fffac7d17f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 03 09:26:44 crc kubenswrapper[4756]: > logger="UnhandledError" Feb 03 09:26:44 crc kubenswrapper[4756]: E0203 09:26:44.163776 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/84bfa317-7260-46d2-8e43-13fffac7d17f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" podUID="84bfa317-7260-46d2-8e43-13fffac7d17f" Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.724230 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.903714 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c","Type":"ContainerStarted","Data":"58dd41e6025ef35e600111e1dc09696fbf0887edce23bf8533300f747637cd5c"} Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.906895 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjjw" event={"ID":"8d6139e2-68a5-4097-b8c2-cf11d11ed241","Type":"ContainerDied","Data":"3186922b7cb46b6157429a9ff4a3ef28305b82cc4e58916cf01fb3d3b2a0c8f8"} Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.906734 4756 generic.go:334] "Generic (PLEG): container finished" podID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerID="3186922b7cb46b6157429a9ff4a3ef28305b82cc4e58916cf01fb3d3b2a0c8f8" exitCode=0 Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.907678 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjjw" event={"ID":"8d6139e2-68a5-4097-b8c2-cf11d11ed241","Type":"ContainerStarted","Data":"9bac9b7120e261034b22c2c262f97eeb4048e09bfae3cbaf5b855823de29a0db"} Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.914849 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bfba7eb1-da32-4bc6-8309-f076ad7b32a0","Type":"ContainerStarted","Data":"dc935433bafd4f7874e97531618336a524f13afe9a5bdc23387dd72162f1de5b"} Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.917475 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb566" event={"ID":"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d","Type":"ContainerStarted","Data":"ed3e3cb7c0abbddb73008fbfe49bd36fdaa6b4b28db55498ed3ca2f41bc6a929"} Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.921826 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" event={"ID":"d1dbab7f-1fdd-49f6-a7a6-904feb481959","Type":"ContainerStarted","Data":"2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48"} Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.924437 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.928727 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95" event={"ID":"15695b6e-11fa-4913-b502-a78974ddf81c","Type":"ContainerStarted","Data":"4e663720f19d06d9aa566aebfa271ba7f5864cf87f93312db6a7e0496d6f095e"} Feb 03 09:26:44 crc kubenswrapper[4756]: I0203 09:26:44.947543 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" podStartSLOduration=6.593444367 podStartE2EDuration="15.947520158s" podCreationTimestamp="2026-02-03 09:26:29 +0000 UTC" firstStartedPulling="2026-02-03 09:26:33.556936144 +0000 UTC m=+984.707403519" lastFinishedPulling="2026-02-03 09:26:42.911011935 +0000 UTC m=+994.061479310" observedRunningTime="2026-02-03 09:26:44.941940501 +0000 UTC m=+996.092407896" watchObservedRunningTime="2026-02-03 09:26:44.947520158 +0000 UTC m=+996.097987533" Feb 03 09:26:45 crc kubenswrapper[4756]: I0203 09:26:45.622892 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bcab1be-ec76-463a-b5e7-f5c8e0f78338" path="/var/lib/kubelet/pods/3bcab1be-ec76-463a-b5e7-f5c8e0f78338/volumes" Feb 03 09:26:45 crc kubenswrapper[4756]: I0203 09:26:45.623286 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b58a001a-2f8b-4dfb-acb5-eecbdd2761ae" path="/var/lib/kubelet/pods/b58a001a-2f8b-4dfb-acb5-eecbdd2761ae/volumes" Feb 03 09:26:45 crc kubenswrapper[4756]: I0203 09:26:45.937683 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a","Type":"ContainerStarted","Data":"240aea3233e6b6a19653e31361142503ce15b09c1d10e2eb4d5ac7bf3b9c80a6"} Feb 03 09:26:49 crc kubenswrapper[4756]: I0203 09:26:49.902086 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:26:49 crc kubenswrapper[4756]: I0203 09:26:49.948678 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9sv9k"] Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.975882 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" event={"ID":"84bfa317-7260-46d2-8e43-13fffac7d17f","Type":"ContainerStarted","Data":"c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7"} Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.976466 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" podUID="84bfa317-7260-46d2-8e43-13fffac7d17f" containerName="dnsmasq-dns" containerID="cri-o://c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7" gracePeriod=10 Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.977038 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.982408 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c","Type":"ContainerStarted","Data":"684c7fdde336ef889eae558a1a1bc56fda86adc1abf41c1c8be79d163f358d58"} Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.984886 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a","Type":"ContainerStarted","Data":"c3fe0d88c9080e41dcc6ecd2aa43e8ceb3657efd8a507dce01c67181b2850f22"} Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.988788 4756 generic.go:334] "Generic (PLEG): container finished" podID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerID="3d1424c47d55ae31a06dfa1e6f18ccb858bb53f6cff0f91e464d120287078d51" exitCode=0 Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.989523 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjjw" event={"ID":"8d6139e2-68a5-4097-b8c2-cf11d11ed241","Type":"ContainerDied","Data":"3d1424c47d55ae31a06dfa1e6f18ccb858bb53f6cff0f91e464d120287078d51"} Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.992197 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bfba7eb1-da32-4bc6-8309-f076ad7b32a0","Type":"ContainerStarted","Data":"243cc71dd2e8e498549172bb835c9f358fee4d41d2f65679f4b995c1d801219f"} Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.994986 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" podStartSLOduration=10.165617141 podStartE2EDuration="22.994969965s" podCreationTimestamp="2026-02-03 09:26:29 +0000 UTC" firstStartedPulling="2026-02-03 09:26:30.075294789 +0000 UTC m=+981.225762164" lastFinishedPulling="2026-02-03 09:26:42.904647603 +0000 UTC m=+994.055114988" observedRunningTime="2026-02-03 09:26:51.991928738 +0000 UTC m=+1003.142396123" watchObservedRunningTime="2026-02-03 09:26:51.994969965 +0000 UTC m=+1003.145437340" Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.995263 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"773ac9a0-544a-4c28-a100-59f10ec789c7","Type":"ContainerStarted","Data":"1791a77fd71211fb49839eac63107ebf2ba24862a135dcb5b1f85769c0591e6d"} Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.995396 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.997210 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ce92a835-f667-4f5b-b9e1-c8afcf476b9a","Type":"ContainerStarted","Data":"e2962440675ad7c84eb6801ff1de2bd02c08c1f944f69152647250554a93bce7"} Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.997307 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 03 09:26:51 crc kubenswrapper[4756]: I0203 09:26:51.999253 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb566" event={"ID":"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d","Type":"ContainerStarted","Data":"c612af4f4a79738aa020b4558a2d97eaefbf651e658d5e9155ea1a500cd90925"} Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.001655 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95" event={"ID":"15695b6e-11fa-4913-b502-a78974ddf81c","Type":"ContainerStarted","Data":"4ef5dff59a24510087633e9944f04ce7ee808e35aa6e1cf4cde94f0a10789391"} Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.002355 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-wnf95" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.003805 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"77baa89d-5260-4b1d-a633-a4f3e37ed6bb","Type":"ContainerStarted","Data":"9d5ef2719f3d096d0fb8187155eab56ae10af4584be2c44cbb2853e5bf92cc60"} Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.053024 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.813664641 podStartE2EDuration="17.053004531s" podCreationTimestamp="2026-02-03 09:26:35 +0000 UTC" firstStartedPulling="2026-02-03 09:26:43.800063197 +0000 UTC m=+994.950530572" lastFinishedPulling="2026-02-03 09:26:51.039403097 +0000 UTC m=+1002.189870462" observedRunningTime="2026-02-03 09:26:52.049639633 +0000 UTC m=+1003.200106998" watchObservedRunningTime="2026-02-03 09:26:52.053004531 +0000 UTC m=+1003.203471906" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.093102 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-wnf95" podStartSLOduration=5.882596503 podStartE2EDuration="13.093078025s" podCreationTimestamp="2026-02-03 09:26:39 +0000 UTC" firstStartedPulling="2026-02-03 09:26:43.828683557 +0000 UTC m=+994.979150932" lastFinishedPulling="2026-02-03 09:26:51.039165079 +0000 UTC m=+1002.189632454" observedRunningTime="2026-02-03 09:26:52.089927035 +0000 UTC m=+1003.240394410" watchObservedRunningTime="2026-02-03 09:26:52.093078025 +0000 UTC m=+1003.243545400" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.133883 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=11.923022849 podStartE2EDuration="19.133866043s" podCreationTimestamp="2026-02-03 09:26:33 +0000 UTC" firstStartedPulling="2026-02-03 09:26:43.76118982 +0000 UTC m=+994.911657195" lastFinishedPulling="2026-02-03 09:26:50.972033014 +0000 UTC m=+1002.122500389" observedRunningTime="2026-02-03 09:26:52.130414673 +0000 UTC m=+1003.280882048" watchObservedRunningTime="2026-02-03 09:26:52.133866043 +0000 UTC m=+1003.284333408" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.412396 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.558836 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkttw\" (UniqueName: \"kubernetes.io/projected/84bfa317-7260-46d2-8e43-13fffac7d17f-kube-api-access-nkttw\") pod \"84bfa317-7260-46d2-8e43-13fffac7d17f\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.558931 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-config\") pod \"84bfa317-7260-46d2-8e43-13fffac7d17f\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.558969 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-dns-svc\") pod \"84bfa317-7260-46d2-8e43-13fffac7d17f\" (UID: \"84bfa317-7260-46d2-8e43-13fffac7d17f\") " Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.565414 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84bfa317-7260-46d2-8e43-13fffac7d17f-kube-api-access-nkttw" (OuterVolumeSpecName: "kube-api-access-nkttw") pod "84bfa317-7260-46d2-8e43-13fffac7d17f" (UID: "84bfa317-7260-46d2-8e43-13fffac7d17f"). InnerVolumeSpecName "kube-api-access-nkttw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.599685 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "84bfa317-7260-46d2-8e43-13fffac7d17f" (UID: "84bfa317-7260-46d2-8e43-13fffac7d17f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.604873 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-config" (OuterVolumeSpecName: "config") pod "84bfa317-7260-46d2-8e43-13fffac7d17f" (UID: "84bfa317-7260-46d2-8e43-13fffac7d17f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.660854 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkttw\" (UniqueName: \"kubernetes.io/projected/84bfa317-7260-46d2-8e43-13fffac7d17f-kube-api-access-nkttw\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.660889 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:52 crc kubenswrapper[4756]: I0203 09:26:52.660902 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84bfa317-7260-46d2-8e43-13fffac7d17f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.012910 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c","Type":"ContainerStarted","Data":"ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f"} Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.015400 4756 generic.go:334] "Generic (PLEG): container finished" podID="84bfa317-7260-46d2-8e43-13fffac7d17f" containerID="c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7" exitCode=0 Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.015458 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.015469 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" event={"ID":"84bfa317-7260-46d2-8e43-13fffac7d17f","Type":"ContainerDied","Data":"c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7"} Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.015513 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9sv9k" event={"ID":"84bfa317-7260-46d2-8e43-13fffac7d17f","Type":"ContainerDied","Data":"3a051ad86d9331b6ac540b142b0881c6eac993427437218be93e3bd1ea2de735"} Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.015533 4756 scope.go:117] "RemoveContainer" containerID="c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.017811 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f2aeafe0-b129-44d5-86d9-0a95a7ac180d","Type":"ContainerStarted","Data":"30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53"} Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.019698 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjjw" event={"ID":"8d6139e2-68a5-4097-b8c2-cf11d11ed241","Type":"ContainerStarted","Data":"614aa6089127a0ef8b7d200a89ab1576ca48c99f2f5455e1152a7dc7808be5c3"} Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.021028 4756 generic.go:334] "Generic (PLEG): container finished" podID="4600dd3a-bdd0-44b1-8a6d-f34313d5df5d" containerID="c612af4f4a79738aa020b4558a2d97eaefbf651e658d5e9155ea1a500cd90925" exitCode=0 Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.021090 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb566" event={"ID":"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d","Type":"ContainerDied","Data":"c612af4f4a79738aa020b4558a2d97eaefbf651e658d5e9155ea1a500cd90925"} Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.097422 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qcjjw" podStartSLOduration=12.641809014 podStartE2EDuration="20.097405004s" podCreationTimestamp="2026-02-03 09:26:33 +0000 UTC" firstStartedPulling="2026-02-03 09:26:44.969153446 +0000 UTC m=+996.119620821" lastFinishedPulling="2026-02-03 09:26:52.424749436 +0000 UTC m=+1003.575216811" observedRunningTime="2026-02-03 09:26:53.090456802 +0000 UTC m=+1004.240924187" watchObservedRunningTime="2026-02-03 09:26:53.097405004 +0000 UTC m=+1004.247872379" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.110653 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9sv9k"] Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.116414 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9sv9k"] Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.315877 4756 scope.go:117] "RemoveContainer" containerID="4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.386856 4756 scope.go:117] "RemoveContainer" containerID="c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7" Feb 03 09:26:53 crc kubenswrapper[4756]: E0203 09:26:53.387309 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7\": container with ID starting with c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7 not found: ID does not exist" containerID="c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.387340 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7"} err="failed to get container status \"c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7\": rpc error: code = NotFound desc = could not find container \"c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7\": container with ID starting with c21b745827f9bfec067f7e4a6a7eda55cf1a32ba0c3f7a5a66a9047d8aa1ddb7 not found: ID does not exist" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.387359 4756 scope.go:117] "RemoveContainer" containerID="4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0" Feb 03 09:26:53 crc kubenswrapper[4756]: E0203 09:26:53.387861 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0\": container with ID starting with 4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0 not found: ID does not exist" containerID="4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.387894 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0"} err="failed to get container status \"4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0\": rpc error: code = NotFound desc = could not find container \"4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0\": container with ID starting with 4b9cd9a7b7761c9d9160f082208b94af477572ff0569040e56429ace4560cac0 not found: ID does not exist" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.624358 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84bfa317-7260-46d2-8e43-13fffac7d17f" path="/var/lib/kubelet/pods/84bfa317-7260-46d2-8e43-13fffac7d17f/volumes" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.782796 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:53 crc kubenswrapper[4756]: I0203 09:26:53.783741 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.031621 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bfba7eb1-da32-4bc6-8309-f076ad7b32a0","Type":"ContainerStarted","Data":"4e27e6ffaa6c1ace0b2ba86088a19d13fe79fd22139d32c2bf48126f7f70b10a"} Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.035164 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb566" event={"ID":"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d","Type":"ContainerStarted","Data":"9ef9b83af6ecbea37a5001cd907bc4756d164fb8b810c08be72d89cbfe38cde8"} Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.035192 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fb566" event={"ID":"4600dd3a-bdd0-44b1-8a6d-f34313d5df5d","Type":"ContainerStarted","Data":"8a3e7c0ad04d1c08b42478baf7f41135c7b4d27cf4dfc11b8cab1f840625be4a"} Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.035354 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.035399 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.040987 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a2da6bae-fd2d-4f9c-b2aa-440d4264f17a","Type":"ContainerStarted","Data":"7d55f3771a4d766277f5c7420000ac20d4ae9a28d69f4a8a75a7ac7d576d2051"} Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.057301 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.677417085 podStartE2EDuration="13.057280588s" podCreationTimestamp="2026-02-03 09:26:41 +0000 UTC" firstStartedPulling="2026-02-03 09:26:44.028796633 +0000 UTC m=+995.179264008" lastFinishedPulling="2026-02-03 09:26:53.408660136 +0000 UTC m=+1004.559127511" observedRunningTime="2026-02-03 09:26:54.053486037 +0000 UTC m=+1005.203953412" watchObservedRunningTime="2026-02-03 09:26:54.057280588 +0000 UTC m=+1005.207747963" Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.075065 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-fb566" podStartSLOduration=7.9386692 podStartE2EDuration="15.075050104s" podCreationTimestamp="2026-02-03 09:26:39 +0000 UTC" firstStartedPulling="2026-02-03 09:26:43.890099921 +0000 UTC m=+995.040567296" lastFinishedPulling="2026-02-03 09:26:51.026480825 +0000 UTC m=+1002.176948200" observedRunningTime="2026-02-03 09:26:54.07400165 +0000 UTC m=+1005.224469025" watchObservedRunningTime="2026-02-03 09:26:54.075050104 +0000 UTC m=+1005.225517479" Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.104019 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=6.7161029370000005 podStartE2EDuration="15.103992884s" podCreationTimestamp="2026-02-03 09:26:39 +0000 UTC" firstStartedPulling="2026-02-03 09:26:45.01233908 +0000 UTC m=+996.162806455" lastFinishedPulling="2026-02-03 09:26:53.400229027 +0000 UTC m=+1004.550696402" observedRunningTime="2026-02-03 09:26:54.099174351 +0000 UTC m=+1005.249641736" watchObservedRunningTime="2026-02-03 09:26:54.103992884 +0000 UTC m=+1005.254460249" Feb 03 09:26:54 crc kubenswrapper[4756]: I0203 09:26:54.828820 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-qcjjw" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="registry-server" probeResult="failure" output=< Feb 03 09:26:54 crc kubenswrapper[4756]: timeout: failed to connect service ":50051" within 1s Feb 03 09:26:54 crc kubenswrapper[4756]: > Feb 03 09:26:55 crc kubenswrapper[4756]: I0203 09:26:55.171245 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:55 crc kubenswrapper[4756]: I0203 09:26:55.211061 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:55 crc kubenswrapper[4756]: I0203 09:26:55.527007 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:55 crc kubenswrapper[4756]: I0203 09:26:55.527052 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:55 crc kubenswrapper[4756]: I0203 09:26:55.559120 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.053254 4756 generic.go:334] "Generic (PLEG): container finished" podID="47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c" containerID="684c7fdde336ef889eae558a1a1bc56fda86adc1abf41c1c8be79d163f358d58" exitCode=0 Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.053347 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c","Type":"ContainerDied","Data":"684c7fdde336ef889eae558a1a1bc56fda86adc1abf41c1c8be79d163f358d58"} Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.054923 4756 generic.go:334] "Generic (PLEG): container finished" podID="77baa89d-5260-4b1d-a633-a4f3e37ed6bb" containerID="9d5ef2719f3d096d0fb8187155eab56ae10af4584be2c44cbb2853e5bf92cc60" exitCode=0 Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.055423 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"77baa89d-5260-4b1d-a633-a4f3e37ed6bb","Type":"ContainerDied","Data":"9d5ef2719f3d096d0fb8187155eab56ae10af4584be2c44cbb2853e5bf92cc60"} Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.055474 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.099238 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.103150 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.117781 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.283672 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qh844"] Feb 03 09:26:56 crc kubenswrapper[4756]: E0203 09:26:56.284011 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84bfa317-7260-46d2-8e43-13fffac7d17f" containerName="init" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.284025 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="84bfa317-7260-46d2-8e43-13fffac7d17f" containerName="init" Feb 03 09:26:56 crc kubenswrapper[4756]: E0203 09:26:56.284035 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84bfa317-7260-46d2-8e43-13fffac7d17f" containerName="dnsmasq-dns" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.284041 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="84bfa317-7260-46d2-8e43-13fffac7d17f" containerName="dnsmasq-dns" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.284180 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="84bfa317-7260-46d2-8e43-13fffac7d17f" containerName="dnsmasq-dns" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.284953 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.287002 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.298304 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qh844"] Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.333162 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-d4749"] Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.334155 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.337130 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.347277 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-d4749"] Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.412994 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qh844"] Feb 03 09:26:56 crc kubenswrapper[4756]: E0203 09:26:56.418790 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-mf548 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" podUID="4b41e138-375b-44ef-bc98-886661434be5" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.419372 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf548\" (UniqueName: \"kubernetes.io/projected/4b41e138-375b-44ef-bc98-886661434be5-kube-api-access-mf548\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.419401 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.419439 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.419473 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-config\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.443016 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-tstcd"] Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.444186 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.451530 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.464692 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tstcd"] Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.524382 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-ovs-rundir\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.524536 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-combined-ca-bundle\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.524656 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.524806 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf548\" (UniqueName: \"kubernetes.io/projected/4b41e138-375b-44ef-bc98-886661434be5-kube-api-access-mf548\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.524843 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.524902 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-ovn-rundir\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.524942 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-config\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.524976 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.525013 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vg8j\" (UniqueName: \"kubernetes.io/projected/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-kube-api-access-8vg8j\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.525054 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-config\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.526184 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.526217 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-config\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.526909 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.546920 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.548366 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.567322 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf548\" (UniqueName: \"kubernetes.io/projected/4b41e138-375b-44ef-bc98-886661434be5-kube-api-access-mf548\") pod \"dnsmasq-dns-5bf47b49b7-qh844\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.568624 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.568731 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.568869 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.569475 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-vwnn2" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.587761 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626315 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vg8j\" (UniqueName: \"kubernetes.io/projected/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-kube-api-access-8vg8j\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626374 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xrl5\" (UniqueName: \"kubernetes.io/projected/c8520b34-412a-4d1c-a610-4c2e54df94bb-kube-api-access-7xrl5\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626439 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-ovs-rundir\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626481 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-dns-svc\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626514 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-combined-ca-bundle\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626548 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626570 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626597 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626658 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-config\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626693 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-ovn-rundir\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.626720 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-config\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.627472 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-config\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.633763 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-ovn-rundir\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.634135 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-ovs-rundir\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.638074 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-combined-ca-bundle\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.640193 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.662008 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vg8j\" (UniqueName: \"kubernetes.io/projected/f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a-kube-api-access-8vg8j\") pod \"ovn-controller-metrics-d4749\" (UID: \"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a\") " pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.728469 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swzh4\" (UniqueName: \"kubernetes.io/projected/912cca3d-0179-4199-b236-a6ae37c9a80d-kube-api-access-swzh4\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.728768 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-dns-svc\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.728892 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.729302 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.730136 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.730783 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/912cca3d-0179-4199-b236-a6ae37c9a80d-scripts\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.730889 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.730743 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.730054 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.729640 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-dns-svc\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.731302 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912cca3d-0179-4199-b236-a6ae37c9a80d-config\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.731380 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.731422 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-config\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.731487 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/912cca3d-0179-4199-b236-a6ae37c9a80d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.731588 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xrl5\" (UniqueName: \"kubernetes.io/projected/c8520b34-412a-4d1c-a610-4c2e54df94bb-kube-api-access-7xrl5\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.732253 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-config\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.748975 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xrl5\" (UniqueName: \"kubernetes.io/projected/c8520b34-412a-4d1c-a610-4c2e54df94bb-kube-api-access-7xrl5\") pod \"dnsmasq-dns-8554648995-tstcd\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.768784 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.833395 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swzh4\" (UniqueName: \"kubernetes.io/projected/912cca3d-0179-4199-b236-a6ae37c9a80d-kube-api-access-swzh4\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.833792 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.833959 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/912cca3d-0179-4199-b236-a6ae37c9a80d-scripts\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.834048 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.834149 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912cca3d-0179-4199-b236-a6ae37c9a80d-config\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.834272 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.834370 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/912cca3d-0179-4199-b236-a6ae37c9a80d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.834780 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/912cca3d-0179-4199-b236-a6ae37c9a80d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.835114 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/912cca3d-0179-4199-b236-a6ae37c9a80d-config\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.835790 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/912cca3d-0179-4199-b236-a6ae37c9a80d-scripts\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.837162 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.837282 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.837343 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/912cca3d-0179-4199-b236-a6ae37c9a80d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.854051 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swzh4\" (UniqueName: \"kubernetes.io/projected/912cca3d-0179-4199-b236-a6ae37c9a80d-kube-api-access-swzh4\") pod \"ovn-northd-0\" (UID: \"912cca3d-0179-4199-b236-a6ae37c9a80d\") " pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.921767 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 03 09:26:56 crc kubenswrapper[4756]: I0203 09:26:56.956612 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-d4749" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.074166 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c","Type":"ContainerStarted","Data":"9a26128b6a8782efcd23612c64566a4cf571b4fba484c708b80a8c0cf808ed97"} Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.082584 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.082718 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"77baa89d-5260-4b1d-a633-a4f3e37ed6bb","Type":"ContainerStarted","Data":"07af6785101171c42d68d33e43ad31f49ad3714202e9ccbc3cef577a5ee9d85a"} Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.129346 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=19.847722088 podStartE2EDuration="27.129330163s" podCreationTimestamp="2026-02-03 09:26:30 +0000 UTC" firstStartedPulling="2026-02-03 09:26:43.748131584 +0000 UTC m=+994.898598959" lastFinishedPulling="2026-02-03 09:26:51.029739659 +0000 UTC m=+1002.180207034" observedRunningTime="2026-02-03 09:26:57.123752096 +0000 UTC m=+1008.274219471" watchObservedRunningTime="2026-02-03 09:26:57.129330163 +0000 UTC m=+1008.279797538" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.129938 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.129998 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=17.891641985 podStartE2EDuration="25.129993314s" podCreationTimestamp="2026-02-03 09:26:32 +0000 UTC" firstStartedPulling="2026-02-03 09:26:43.830297628 +0000 UTC m=+994.980765003" lastFinishedPulling="2026-02-03 09:26:51.068648967 +0000 UTC m=+1002.219116332" observedRunningTime="2026-02-03 09:26:57.098072779 +0000 UTC m=+1008.248540154" watchObservedRunningTime="2026-02-03 09:26:57.129993314 +0000 UTC m=+1008.280460689" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.208751 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tstcd"] Feb 03 09:26:57 crc kubenswrapper[4756]: W0203 09:26:57.212967 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8520b34_412a_4d1c_a610_4c2e54df94bb.slice/crio-f4ea14aba78af2a7c83e85cdb8b56b39f211bff390859d579aed80c502d24db4 WatchSource:0}: Error finding container f4ea14aba78af2a7c83e85cdb8b56b39f211bff390859d579aed80c502d24db4: Status 404 returned error can't find the container with id f4ea14aba78af2a7c83e85cdb8b56b39f211bff390859d579aed80c502d24db4 Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.244913 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf548\" (UniqueName: \"kubernetes.io/projected/4b41e138-375b-44ef-bc98-886661434be5-kube-api-access-mf548\") pod \"4b41e138-375b-44ef-bc98-886661434be5\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.245001 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-config\") pod \"4b41e138-375b-44ef-bc98-886661434be5\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.245043 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-dns-svc\") pod \"4b41e138-375b-44ef-bc98-886661434be5\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.245168 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-ovsdbserver-nb\") pod \"4b41e138-375b-44ef-bc98-886661434be5\" (UID: \"4b41e138-375b-44ef-bc98-886661434be5\") " Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.246077 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4b41e138-375b-44ef-bc98-886661434be5" (UID: "4b41e138-375b-44ef-bc98-886661434be5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.246149 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4b41e138-375b-44ef-bc98-886661434be5" (UID: "4b41e138-375b-44ef-bc98-886661434be5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.247509 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-config" (OuterVolumeSpecName: "config") pod "4b41e138-375b-44ef-bc98-886661434be5" (UID: "4b41e138-375b-44ef-bc98-886661434be5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.249668 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b41e138-375b-44ef-bc98-886661434be5-kube-api-access-mf548" (OuterVolumeSpecName: "kube-api-access-mf548") pod "4b41e138-375b-44ef-bc98-886661434be5" (UID: "4b41e138-375b-44ef-bc98-886661434be5"). InnerVolumeSpecName "kube-api-access-mf548". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.347739 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.347770 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.347779 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b41e138-375b-44ef-bc98-886661434be5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.347790 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf548\" (UniqueName: \"kubernetes.io/projected/4b41e138-375b-44ef-bc98-886661434be5-kube-api-access-mf548\") on node \"crc\" DevicePath \"\"" Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.360755 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 03 09:26:57 crc kubenswrapper[4756]: W0203 09:26:57.371822 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod912cca3d_0179_4199_b236_a6ae37c9a80d.slice/crio-adfcdfe00a565d86ce6122b7a93779314ec4f1d2f05f51753e63ddaa7c946560 WatchSource:0}: Error finding container adfcdfe00a565d86ce6122b7a93779314ec4f1d2f05f51753e63ddaa7c946560: Status 404 returned error can't find the container with id adfcdfe00a565d86ce6122b7a93779314ec4f1d2f05f51753e63ddaa7c946560 Feb 03 09:26:57 crc kubenswrapper[4756]: I0203 09:26:57.455555 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-d4749"] Feb 03 09:26:57 crc kubenswrapper[4756]: W0203 09:26:57.459492 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6c5f3c6_ec4f_4c2a_beb0_128e53ecb89a.slice/crio-4f66131659fff3aa687a06736d9f81af78f8d0543232a270ad37c91abf90d389 WatchSource:0}: Error finding container 4f66131659fff3aa687a06736d9f81af78f8d0543232a270ad37c91abf90d389: Status 404 returned error can't find the container with id 4f66131659fff3aa687a06736d9f81af78f8d0543232a270ad37c91abf90d389 Feb 03 09:26:58 crc kubenswrapper[4756]: I0203 09:26:58.095815 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tstcd" event={"ID":"c8520b34-412a-4d1c-a610-4c2e54df94bb","Type":"ContainerStarted","Data":"f4ea14aba78af2a7c83e85cdb8b56b39f211bff390859d579aed80c502d24db4"} Feb 03 09:26:58 crc kubenswrapper[4756]: I0203 09:26:58.099490 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"912cca3d-0179-4199-b236-a6ae37c9a80d","Type":"ContainerStarted","Data":"adfcdfe00a565d86ce6122b7a93779314ec4f1d2f05f51753e63ddaa7c946560"} Feb 03 09:26:58 crc kubenswrapper[4756]: I0203 09:26:58.101665 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-d4749" event={"ID":"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a","Type":"ContainerStarted","Data":"4f66131659fff3aa687a06736d9f81af78f8d0543232a270ad37c91abf90d389"} Feb 03 09:26:58 crc kubenswrapper[4756]: I0203 09:26:58.101754 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-qh844" Feb 03 09:26:58 crc kubenswrapper[4756]: I0203 09:26:58.143591 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qh844"] Feb 03 09:26:58 crc kubenswrapper[4756]: I0203 09:26:58.151159 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qh844"] Feb 03 09:26:59 crc kubenswrapper[4756]: I0203 09:26:59.111150 4756 generic.go:334] "Generic (PLEG): container finished" podID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerID="46ba9e9b73efcea49e37d1e0eb409691cd1bf0bf7137fdba5d1b6008ed40a062" exitCode=0 Feb 03 09:26:59 crc kubenswrapper[4756]: I0203 09:26:59.111253 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tstcd" event={"ID":"c8520b34-412a-4d1c-a610-4c2e54df94bb","Type":"ContainerDied","Data":"46ba9e9b73efcea49e37d1e0eb409691cd1bf0bf7137fdba5d1b6008ed40a062"} Feb 03 09:26:59 crc kubenswrapper[4756]: I0203 09:26:59.113059 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-d4749" event={"ID":"f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a","Type":"ContainerStarted","Data":"39eeabd5c267065975cb521e99ab044ce0110421a0c9a24ea0c6c73349682294"} Feb 03 09:26:59 crc kubenswrapper[4756]: I0203 09:26:59.140623 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 03 09:26:59 crc kubenswrapper[4756]: I0203 09:26:59.181345 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-d4749" podStartSLOduration=3.181327869 podStartE2EDuration="3.181327869s" podCreationTimestamp="2026-02-03 09:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:26:59.154217996 +0000 UTC m=+1010.304685371" watchObservedRunningTime="2026-02-03 09:26:59.181327869 +0000 UTC m=+1010.331795244" Feb 03 09:26:59 crc kubenswrapper[4756]: I0203 09:26:59.626813 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b41e138-375b-44ef-bc98-886661434be5" path="/var/lib/kubelet/pods/4b41e138-375b-44ef-bc98-886661434be5/volumes" Feb 03 09:27:00 crc kubenswrapper[4756]: I0203 09:27:00.121570 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tstcd" event={"ID":"c8520b34-412a-4d1c-a610-4c2e54df94bb","Type":"ContainerStarted","Data":"e797442bdd5883f8f7f0b4d8bd804d0ca6544b5f91950be1c3af5b089b6f6d68"} Feb 03 09:27:00 crc kubenswrapper[4756]: I0203 09:27:00.122486 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:27:00 crc kubenswrapper[4756]: I0203 09:27:00.124614 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"912cca3d-0179-4199-b236-a6ae37c9a80d","Type":"ContainerStarted","Data":"4c5724bf898381d97a9dc567e5985b439e9fa14b6ab6cdd545fdf244f813bc82"} Feb 03 09:27:00 crc kubenswrapper[4756]: I0203 09:27:00.124664 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"912cca3d-0179-4199-b236-a6ae37c9a80d","Type":"ContainerStarted","Data":"e1b507f9b8e641508c35dc020f12c9d9913dacd92d7fe506a6b70be9fb1181cc"} Feb 03 09:27:00 crc kubenswrapper[4756]: I0203 09:27:00.124809 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 03 09:27:00 crc kubenswrapper[4756]: I0203 09:27:00.195047 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-tstcd" podStartSLOduration=4.195023665 podStartE2EDuration="4.195023665s" podCreationTimestamp="2026-02-03 09:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:00.194502459 +0000 UTC m=+1011.344969834" watchObservedRunningTime="2026-02-03 09:27:00.195023665 +0000 UTC m=+1011.345491040" Feb 03 09:27:00 crc kubenswrapper[4756]: I0203 09:27:00.214679 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.069088337 podStartE2EDuration="4.21466013s" podCreationTimestamp="2026-02-03 09:26:56 +0000 UTC" firstStartedPulling="2026-02-03 09:26:57.373174879 +0000 UTC m=+1008.523642254" lastFinishedPulling="2026-02-03 09:26:59.518746682 +0000 UTC m=+1010.669214047" observedRunningTime="2026-02-03 09:27:00.209556337 +0000 UTC m=+1011.360023712" watchObservedRunningTime="2026-02-03 09:27:00.21466013 +0000 UTC m=+1011.365127505" Feb 03 09:27:02 crc kubenswrapper[4756]: I0203 09:27:02.325088 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 03 09:27:02 crc kubenswrapper[4756]: I0203 09:27:02.325427 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 03 09:27:02 crc kubenswrapper[4756]: I0203 09:27:02.395919 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.212305 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.737067 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.737376 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.750544 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-aaef-account-create-update-cjsjc"] Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.757646 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-aaef-account-create-update-cjsjc"] Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.757755 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.761989 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.799329 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-lzqmn"] Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.800297 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.806312 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lzqmn"] Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.859814 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-operator-scripts\") pod \"keystone-aaef-account-create-update-cjsjc\" (UID: \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\") " pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.859923 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r446\" (UniqueName: \"kubernetes.io/projected/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-kube-api-access-6r446\") pod \"keystone-aaef-account-create-update-cjsjc\" (UID: \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\") " pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.961740 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344e964e-c1ec-4fdc-82a3-223389b11f1a-operator-scripts\") pod \"keystone-db-create-lzqmn\" (UID: \"344e964e-c1ec-4fdc-82a3-223389b11f1a\") " pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.961850 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r446\" (UniqueName: \"kubernetes.io/projected/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-kube-api-access-6r446\") pod \"keystone-aaef-account-create-update-cjsjc\" (UID: \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\") " pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.961947 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl9jf\" (UniqueName: \"kubernetes.io/projected/344e964e-c1ec-4fdc-82a3-223389b11f1a-kube-api-access-kl9jf\") pod \"keystone-db-create-lzqmn\" (UID: \"344e964e-c1ec-4fdc-82a3-223389b11f1a\") " pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.962003 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-operator-scripts\") pod \"keystone-aaef-account-create-update-cjsjc\" (UID: \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\") " pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.962947 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-operator-scripts\") pod \"keystone-aaef-account-create-update-cjsjc\" (UID: \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\") " pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:03 crc kubenswrapper[4756]: I0203 09:27:03.990082 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r446\" (UniqueName: \"kubernetes.io/projected/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-kube-api-access-6r446\") pod \"keystone-aaef-account-create-update-cjsjc\" (UID: \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\") " pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.018830 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.021942 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-sl9vg"] Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.023744 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.047034 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sl9vg"] Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.063783 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl9jf\" (UniqueName: \"kubernetes.io/projected/344e964e-c1ec-4fdc-82a3-223389b11f1a-kube-api-access-kl9jf\") pod \"keystone-db-create-lzqmn\" (UID: \"344e964e-c1ec-4fdc-82a3-223389b11f1a\") " pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.063889 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344e964e-c1ec-4fdc-82a3-223389b11f1a-operator-scripts\") pod \"keystone-db-create-lzqmn\" (UID: \"344e964e-c1ec-4fdc-82a3-223389b11f1a\") " pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.064887 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344e964e-c1ec-4fdc-82a3-223389b11f1a-operator-scripts\") pod \"keystone-db-create-lzqmn\" (UID: \"344e964e-c1ec-4fdc-82a3-223389b11f1a\") " pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.081040 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.097769 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.107155 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-187d-account-create-update-8rkgm"] Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.108707 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.113826 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.123260 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl9jf\" (UniqueName: \"kubernetes.io/projected/344e964e-c1ec-4fdc-82a3-223389b11f1a-kube-api-access-kl9jf\") pod \"keystone-db-create-lzqmn\" (UID: \"344e964e-c1ec-4fdc-82a3-223389b11f1a\") " pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.139575 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-187d-account-create-update-8rkgm"] Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.165106 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d909b510-1424-40ce-ad72-0b46a5304b42-operator-scripts\") pod \"placement-db-create-sl9vg\" (UID: \"d909b510-1424-40ce-ad72-0b46a5304b42\") " pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.165262 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq7kh\" (UniqueName: \"kubernetes.io/projected/d909b510-1424-40ce-ad72-0b46a5304b42-kube-api-access-kq7kh\") pod \"placement-db-create-sl9vg\" (UID: \"d909b510-1424-40ce-ad72-0b46a5304b42\") " pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.266989 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d909b510-1424-40ce-ad72-0b46a5304b42-operator-scripts\") pod \"placement-db-create-sl9vg\" (UID: \"d909b510-1424-40ce-ad72-0b46a5304b42\") " pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.267511 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq7kh\" (UniqueName: \"kubernetes.io/projected/d909b510-1424-40ce-ad72-0b46a5304b42-kube-api-access-kq7kh\") pod \"placement-db-create-sl9vg\" (UID: \"d909b510-1424-40ce-ad72-0b46a5304b42\") " pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.267542 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99vh4\" (UniqueName: \"kubernetes.io/projected/dc25ed95-beda-41da-b65f-10dea4ee003f-kube-api-access-99vh4\") pod \"placement-187d-account-create-update-8rkgm\" (UID: \"dc25ed95-beda-41da-b65f-10dea4ee003f\") " pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.267635 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc25ed95-beda-41da-b65f-10dea4ee003f-operator-scripts\") pod \"placement-187d-account-create-update-8rkgm\" (UID: \"dc25ed95-beda-41da-b65f-10dea4ee003f\") " pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.267870 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d909b510-1424-40ce-ad72-0b46a5304b42-operator-scripts\") pod \"placement-db-create-sl9vg\" (UID: \"d909b510-1424-40ce-ad72-0b46a5304b42\") " pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.294808 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq7kh\" (UniqueName: \"kubernetes.io/projected/d909b510-1424-40ce-ad72-0b46a5304b42-kube-api-access-kq7kh\") pod \"placement-db-create-sl9vg\" (UID: \"d909b510-1424-40ce-ad72-0b46a5304b42\") " pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.353723 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.370752 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc25ed95-beda-41da-b65f-10dea4ee003f-operator-scripts\") pod \"placement-187d-account-create-update-8rkgm\" (UID: \"dc25ed95-beda-41da-b65f-10dea4ee003f\") " pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.370862 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99vh4\" (UniqueName: \"kubernetes.io/projected/dc25ed95-beda-41da-b65f-10dea4ee003f-kube-api-access-99vh4\") pod \"placement-187d-account-create-update-8rkgm\" (UID: \"dc25ed95-beda-41da-b65f-10dea4ee003f\") " pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.372147 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc25ed95-beda-41da-b65f-10dea4ee003f-operator-scripts\") pod \"placement-187d-account-create-update-8rkgm\" (UID: \"dc25ed95-beda-41da-b65f-10dea4ee003f\") " pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.394756 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99vh4\" (UniqueName: \"kubernetes.io/projected/dc25ed95-beda-41da-b65f-10dea4ee003f-kube-api-access-99vh4\") pod \"placement-187d-account-create-update-8rkgm\" (UID: \"dc25ed95-beda-41da-b65f-10dea4ee003f\") " pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.416852 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.473887 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.568211 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-aaef-account-create-update-cjsjc"] Feb 03 09:27:04 crc kubenswrapper[4756]: W0203 09:27:04.590675 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d05d9f3_e465_481b_92e0_c4d2c8c640bb.slice/crio-650eeea123d9a59a46e3dcd5c7cbca1a4b7f616a3c40d3cf5ab25ee7ba9f2807 WatchSource:0}: Error finding container 650eeea123d9a59a46e3dcd5c7cbca1a4b7f616a3c40d3cf5ab25ee7ba9f2807: Status 404 returned error can't find the container with id 650eeea123d9a59a46e3dcd5c7cbca1a4b7f616a3c40d3cf5ab25ee7ba9f2807 Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.648510 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjjw"] Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.827641 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sl9vg"] Feb 03 09:27:04 crc kubenswrapper[4756]: W0203 09:27:04.828205 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd909b510_1424_40ce_ad72_0b46a5304b42.slice/crio-3a710add519df44493a6a7c400e1b208108fa11bc2fc6240e9e83a8806c3f99d WatchSource:0}: Error finding container 3a710add519df44493a6a7c400e1b208108fa11bc2fc6240e9e83a8806c3f99d: Status 404 returned error can't find the container with id 3a710add519df44493a6a7c400e1b208108fa11bc2fc6240e9e83a8806c3f99d Feb 03 09:27:04 crc kubenswrapper[4756]: I0203 09:27:04.939860 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lzqmn"] Feb 03 09:27:05 crc kubenswrapper[4756]: I0203 09:27:05.025196 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-187d-account-create-update-8rkgm"] Feb 03 09:27:05 crc kubenswrapper[4756]: W0203 09:27:05.046168 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc25ed95_beda_41da_b65f_10dea4ee003f.slice/crio-1a85f6aedb1472a9fcd37280e66a3d1e1a4e5d21cde07b41a8b9e6c82872cc9d WatchSource:0}: Error finding container 1a85f6aedb1472a9fcd37280e66a3d1e1a4e5d21cde07b41a8b9e6c82872cc9d: Status 404 returned error can't find the container with id 1a85f6aedb1472a9fcd37280e66a3d1e1a4e5d21cde07b41a8b9e6c82872cc9d Feb 03 09:27:05 crc kubenswrapper[4756]: I0203 09:27:05.162952 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-187d-account-create-update-8rkgm" event={"ID":"dc25ed95-beda-41da-b65f-10dea4ee003f","Type":"ContainerStarted","Data":"1a85f6aedb1472a9fcd37280e66a3d1e1a4e5d21cde07b41a8b9e6c82872cc9d"} Feb 03 09:27:05 crc kubenswrapper[4756]: I0203 09:27:05.164211 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lzqmn" event={"ID":"344e964e-c1ec-4fdc-82a3-223389b11f1a","Type":"ContainerStarted","Data":"8ec5597378e0d0c1d0dc501db8a5cb9c34db79e7d9ee6ce6ace3f0700209f013"} Feb 03 09:27:05 crc kubenswrapper[4756]: I0203 09:27:05.167564 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sl9vg" event={"ID":"d909b510-1424-40ce-ad72-0b46a5304b42","Type":"ContainerStarted","Data":"3a710add519df44493a6a7c400e1b208108fa11bc2fc6240e9e83a8806c3f99d"} Feb 03 09:27:05 crc kubenswrapper[4756]: I0203 09:27:05.168834 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-aaef-account-create-update-cjsjc" event={"ID":"4d05d9f3-e465-481b-92e0-c4d2c8c640bb","Type":"ContainerStarted","Data":"650eeea123d9a59a46e3dcd5c7cbca1a4b7f616a3c40d3cf5ab25ee7ba9f2807"} Feb 03 09:27:05 crc kubenswrapper[4756]: I0203 09:27:05.169013 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qcjjw" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="registry-server" containerID="cri-o://614aa6089127a0ef8b7d200a89ab1576ca48c99f2f5455e1152a7dc7808be5c3" gracePeriod=2 Feb 03 09:27:05 crc kubenswrapper[4756]: I0203 09:27:05.435951 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 03 09:27:05 crc kubenswrapper[4756]: I0203 09:27:05.509314 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c" containerName="galera" probeResult="failure" output=< Feb 03 09:27:05 crc kubenswrapper[4756]: wsrep_local_state_comment (Joined) differs from Synced Feb 03 09:27:05 crc kubenswrapper[4756]: > Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.055848 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tstcd"] Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.057073 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-tstcd" podUID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerName="dnsmasq-dns" containerID="cri-o://e797442bdd5883f8f7f0b4d8bd804d0ca6544b5f91950be1c3af5b089b6f6d68" gracePeriod=10 Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.057619 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.134265 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dds6r"] Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.135516 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.153054 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dds6r"] Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.204898 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.204952 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pls4\" (UniqueName: \"kubernetes.io/projected/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-kube-api-access-8pls4\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.204982 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.205012 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-config\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.205062 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.221653 4756 generic.go:334] "Generic (PLEG): container finished" podID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerID="614aa6089127a0ef8b7d200a89ab1576ca48c99f2f5455e1152a7dc7808be5c3" exitCode=0 Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.221877 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjjw" event={"ID":"8d6139e2-68a5-4097-b8c2-cf11d11ed241","Type":"ContainerDied","Data":"614aa6089127a0ef8b7d200a89ab1576ca48c99f2f5455e1152a7dc7808be5c3"} Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.309231 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.309331 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.309361 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pls4\" (UniqueName: \"kubernetes.io/projected/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-kube-api-access-8pls4\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.309381 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.309409 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-config\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.310330 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-config\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.310837 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.311163 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.311521 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.348950 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pls4\" (UniqueName: \"kubernetes.io/projected/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-kube-api-access-8pls4\") pod \"dnsmasq-dns-b8fbc5445-dds6r\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.467077 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.770657 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-tstcd" podUID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: connect: connection refused" Feb 03 09:27:06 crc kubenswrapper[4756]: I0203 09:27:06.879395 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dds6r"] Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.229059 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-aaef-account-create-update-cjsjc" event={"ID":"4d05d9f3-e465-481b-92e0-c4d2c8c640bb","Type":"ContainerStarted","Data":"9d50aed8cb96d489fc52415db103e11db5ee939cc96736b7234b76e4adfa48ef"} Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.230332 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" event={"ID":"1400dee3-79cd-4c19-b349-59c3d1b0bc5c","Type":"ContainerStarted","Data":"fa3c255e624fd63e2d605b897ddd977597db84881449402008a48d2ea3944e29"} Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.231911 4756 generic.go:334] "Generic (PLEG): container finished" podID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerID="e797442bdd5883f8f7f0b4d8bd804d0ca6544b5f91950be1c3af5b089b6f6d68" exitCode=0 Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.231954 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tstcd" event={"ID":"c8520b34-412a-4d1c-a610-4c2e54df94bb","Type":"ContainerDied","Data":"e797442bdd5883f8f7f0b4d8bd804d0ca6544b5f91950be1c3af5b089b6f6d68"} Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.296672 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.303489 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.306798 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.307542 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.307928 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.308029 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-cfvpk" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.318826 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.360427 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.447918 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-sb\") pod \"c8520b34-412a-4d1c-a610-4c2e54df94bb\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448064 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-config\") pod \"c8520b34-412a-4d1c-a610-4c2e54df94bb\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448095 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xrl5\" (UniqueName: \"kubernetes.io/projected/c8520b34-412a-4d1c-a610-4c2e54df94bb-kube-api-access-7xrl5\") pod \"c8520b34-412a-4d1c-a610-4c2e54df94bb\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448156 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-nb\") pod \"c8520b34-412a-4d1c-a610-4c2e54df94bb\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448183 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-dns-svc\") pod \"c8520b34-412a-4d1c-a610-4c2e54df94bb\" (UID: \"c8520b34-412a-4d1c-a610-4c2e54df94bb\") " Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448422 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-cache\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448527 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448603 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-lock\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448663 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448695 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7gkv\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-kube-api-access-w7gkv\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.448726 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.453120 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8520b34-412a-4d1c-a610-4c2e54df94bb-kube-api-access-7xrl5" (OuterVolumeSpecName: "kube-api-access-7xrl5") pod "c8520b34-412a-4d1c-a610-4c2e54df94bb" (UID: "c8520b34-412a-4d1c-a610-4c2e54df94bb"). InnerVolumeSpecName "kube-api-access-7xrl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.488427 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c8520b34-412a-4d1c-a610-4c2e54df94bb" (UID: "c8520b34-412a-4d1c-a610-4c2e54df94bb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.488466 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c8520b34-412a-4d1c-a610-4c2e54df94bb" (UID: "c8520b34-412a-4d1c-a610-4c2e54df94bb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.493421 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c8520b34-412a-4d1c-a610-4c2e54df94bb" (UID: "c8520b34-412a-4d1c-a610-4c2e54df94bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.501532 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-config" (OuterVolumeSpecName: "config") pod "c8520b34-412a-4d1c-a610-4c2e54df94bb" (UID: "c8520b34-412a-4d1c-a610-4c2e54df94bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.525054 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-pqfmb"] Feb 03 09:27:07 crc kubenswrapper[4756]: E0203 09:27:07.525713 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerName="init" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.525729 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerName="init" Feb 03 09:27:07 crc kubenswrapper[4756]: E0203 09:27:07.525763 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerName="dnsmasq-dns" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.525772 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerName="dnsmasq-dns" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.525920 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8520b34-412a-4d1c-a610-4c2e54df94bb" containerName="dnsmasq-dns" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.526409 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.528118 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.529251 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.529276 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.536855 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-pqfmb"] Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.550681 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.550752 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7gkv\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-kube-api-access-w7gkv\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.550795 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.550861 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-cache\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.550883 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.550920 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-lock\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.551001 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.551046 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.551060 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xrl5\" (UniqueName: \"kubernetes.io/projected/c8520b34-412a-4d1c-a610-4c2e54df94bb-kube-api-access-7xrl5\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.551072 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.551083 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8520b34-412a-4d1c-a610-4c2e54df94bb-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.551545 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-lock\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.551842 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: E0203 09:27:07.552704 4756 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 03 09:27:07 crc kubenswrapper[4756]: E0203 09:27:07.552741 4756 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 03 09:27:07 crc kubenswrapper[4756]: E0203 09:27:07.552795 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift podName:d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9 nodeName:}" failed. No retries permitted until 2026-02-03 09:27:08.052776192 +0000 UTC m=+1019.203243567 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift") pod "swift-storage-0" (UID: "d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9") : configmap "swift-ring-files" not found Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.553407 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-cache\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.557648 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.569784 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7gkv\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-kube-api-access-w7gkv\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.580187 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.652424 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-dispersionconf\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.652567 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-etc-swift\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.652656 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l52z\" (UniqueName: \"kubernetes.io/projected/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-kube-api-access-7l52z\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.652724 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-scripts\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.652796 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-swiftconf\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.652919 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-combined-ca-bundle\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.653106 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-ring-data-devices\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.754896 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-ring-data-devices\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.754962 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-dispersionconf\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.754997 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-etc-swift\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.755035 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l52z\" (UniqueName: \"kubernetes.io/projected/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-kube-api-access-7l52z\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.755066 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-scripts\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.755085 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-swiftconf\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.755128 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-combined-ca-bundle\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.755963 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-etc-swift\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.756272 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-ring-data-devices\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.756301 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-scripts\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.758218 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-dispersionconf\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.758961 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-combined-ca-bundle\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.760217 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-swiftconf\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.773090 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l52z\" (UniqueName: \"kubernetes.io/projected/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-kube-api-access-7l52z\") pod \"swift-ring-rebalance-pqfmb\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:07 crc kubenswrapper[4756]: I0203 09:27:07.869593 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.062385 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:08 crc kubenswrapper[4756]: E0203 09:27:08.062685 4756 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 03 09:27:08 crc kubenswrapper[4756]: E0203 09:27:08.062753 4756 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 03 09:27:08 crc kubenswrapper[4756]: E0203 09:27:08.062810 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift podName:d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9 nodeName:}" failed. No retries permitted until 2026-02-03 09:27:09.062793006 +0000 UTC m=+1020.213260381 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift") pod "swift-storage-0" (UID: "d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9") : configmap "swift-ring-files" not found Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.209514 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.246263 4756 generic.go:334] "Generic (PLEG): container finished" podID="4d05d9f3-e465-481b-92e0-c4d2c8c640bb" containerID="9d50aed8cb96d489fc52415db103e11db5ee939cc96736b7234b76e4adfa48ef" exitCode=0 Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.246351 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-aaef-account-create-update-cjsjc" event={"ID":"4d05d9f3-e465-481b-92e0-c4d2c8c640bb","Type":"ContainerDied","Data":"9d50aed8cb96d489fc52415db103e11db5ee939cc96736b7234b76e4adfa48ef"} Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.249918 4756 generic.go:334] "Generic (PLEG): container finished" podID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" containerID="64cd5c06f0373d780a3502feb93099b27653ad1c254c0232dde1169509822c46" exitCode=0 Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.249998 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" event={"ID":"1400dee3-79cd-4c19-b349-59c3d1b0bc5c","Type":"ContainerDied","Data":"64cd5c06f0373d780a3502feb93099b27653ad1c254c0232dde1169509822c46"} Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.253831 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcjjw" event={"ID":"8d6139e2-68a5-4097-b8c2-cf11d11ed241","Type":"ContainerDied","Data":"9bac9b7120e261034b22c2c262f97eeb4048e09bfae3cbaf5b855823de29a0db"} Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.253852 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcjjw" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.253890 4756 scope.go:117] "RemoveContainer" containerID="614aa6089127a0ef8b7d200a89ab1576ca48c99f2f5455e1152a7dc7808be5c3" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.255700 4756 generic.go:334] "Generic (PLEG): container finished" podID="dc25ed95-beda-41da-b65f-10dea4ee003f" containerID="6c943329f74b1bdcce52b7e8805fa9b7f41ab6f68b547aac834cb754b75b62c6" exitCode=0 Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.255870 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-187d-account-create-update-8rkgm" event={"ID":"dc25ed95-beda-41da-b65f-10dea4ee003f","Type":"ContainerDied","Data":"6c943329f74b1bdcce52b7e8805fa9b7f41ab6f68b547aac834cb754b75b62c6"} Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.263242 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-tstcd" event={"ID":"c8520b34-412a-4d1c-a610-4c2e54df94bb","Type":"ContainerDied","Data":"f4ea14aba78af2a7c83e85cdb8b56b39f211bff390859d579aed80c502d24db4"} Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.263290 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-tstcd" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.266847 4756 generic.go:334] "Generic (PLEG): container finished" podID="344e964e-c1ec-4fdc-82a3-223389b11f1a" containerID="0c4c1352d1710a022c86866da9b6b5f4cae10388c10fde9d55f34c5985e208ec" exitCode=0 Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.266928 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lzqmn" event={"ID":"344e964e-c1ec-4fdc-82a3-223389b11f1a","Type":"ContainerDied","Data":"0c4c1352d1710a022c86866da9b6b5f4cae10388c10fde9d55f34c5985e208ec"} Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.272916 4756 generic.go:334] "Generic (PLEG): container finished" podID="d909b510-1424-40ce-ad72-0b46a5304b42" containerID="c71513e6b7ef7e9c0014b955513855c12ac490eb1d6cf98839ffd582b23fd90d" exitCode=0 Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.272964 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sl9vg" event={"ID":"d909b510-1424-40ce-ad72-0b46a5304b42","Type":"ContainerDied","Data":"c71513e6b7ef7e9c0014b955513855c12ac490eb1d6cf98839ffd582b23fd90d"} Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.362926 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-pqfmb"] Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.367756 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mml57\" (UniqueName: \"kubernetes.io/projected/8d6139e2-68a5-4097-b8c2-cf11d11ed241-kube-api-access-mml57\") pod \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.367818 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-catalog-content\") pod \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.367915 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-utilities\") pod \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\" (UID: \"8d6139e2-68a5-4097-b8c2-cf11d11ed241\") " Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.370949 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-utilities" (OuterVolumeSpecName: "utilities") pod "8d6139e2-68a5-4097-b8c2-cf11d11ed241" (UID: "8d6139e2-68a5-4097-b8c2-cf11d11ed241"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.372638 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d6139e2-68a5-4097-b8c2-cf11d11ed241-kube-api-access-mml57" (OuterVolumeSpecName: "kube-api-access-mml57") pod "8d6139e2-68a5-4097-b8c2-cf11d11ed241" (UID: "8d6139e2-68a5-4097-b8c2-cf11d11ed241"). InnerVolumeSpecName "kube-api-access-mml57". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.394155 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d6139e2-68a5-4097-b8c2-cf11d11ed241" (UID: "8d6139e2-68a5-4097-b8c2-cf11d11ed241"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.417490 4756 scope.go:117] "RemoveContainer" containerID="3d1424c47d55ae31a06dfa1e6f18ccb858bb53f6cff0f91e464d120287078d51" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.437701 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tstcd"] Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.444216 4756 scope.go:117] "RemoveContainer" containerID="3186922b7cb46b6157429a9ff4a3ef28305b82cc4e58916cf01fb3d3b2a0c8f8" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.444817 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-tstcd"] Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.464624 4756 scope.go:117] "RemoveContainer" containerID="e797442bdd5883f8f7f0b4d8bd804d0ca6544b5f91950be1c3af5b089b6f6d68" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.470722 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.470754 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6139e2-68a5-4097-b8c2-cf11d11ed241-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.470768 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mml57\" (UniqueName: \"kubernetes.io/projected/8d6139e2-68a5-4097-b8c2-cf11d11ed241-kube-api-access-mml57\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.480664 4756 scope.go:117] "RemoveContainer" containerID="46ba9e9b73efcea49e37d1e0eb409691cd1bf0bf7137fdba5d1b6008ed40a062" Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.608704 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjjw"] Feb 03 09:27:08 crc kubenswrapper[4756]: I0203 09:27:08.619018 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcjjw"] Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.083337 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:09 crc kubenswrapper[4756]: E0203 09:27:09.083530 4756 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 03 09:27:09 crc kubenswrapper[4756]: E0203 09:27:09.083764 4756 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 03 09:27:09 crc kubenswrapper[4756]: E0203 09:27:09.083825 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift podName:d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9 nodeName:}" failed. No retries permitted until 2026-02-03 09:27:11.083804355 +0000 UTC m=+1022.234271730 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift") pod "swift-storage-0" (UID: "d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9") : configmap "swift-ring-files" not found Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.285930 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pqfmb" event={"ID":"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e","Type":"ContainerStarted","Data":"56bab4967dbd539cf2aa6bcc439d6eb2b6595bdc07ceb312bd2fee6e6ccd131c"} Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.293210 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-kdsms"] Feb 03 09:27:09 crc kubenswrapper[4756]: E0203 09:27:09.294365 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="registry-server" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.294390 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="registry-server" Feb 03 09:27:09 crc kubenswrapper[4756]: E0203 09:27:09.294403 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="extract-content" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.294409 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="extract-content" Feb 03 09:27:09 crc kubenswrapper[4756]: E0203 09:27:09.294435 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="extract-utilities" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.294460 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="extract-utilities" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.294661 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" containerName="registry-server" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.295407 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kdsms" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.295319 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" event={"ID":"1400dee3-79cd-4c19-b349-59c3d1b0bc5c","Type":"ContainerStarted","Data":"d9b900cb37e1258bc3cb26422229a731e02bef084ec5edcfe634acf9aca0b526"} Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.295917 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.302814 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kdsms"] Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.320159 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" podStartSLOduration=3.320136304 podStartE2EDuration="3.320136304s" podCreationTimestamp="2026-02-03 09:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:09.31406056 +0000 UTC m=+1020.464527935" watchObservedRunningTime="2026-02-03 09:27:09.320136304 +0000 UTC m=+1020.470603679" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.378018 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-60b6-account-create-update-xwbqp"] Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.379194 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.381255 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.387918 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a643eeaf-1782-49f9-bbb7-55bdd019f68e-operator-scripts\") pod \"glance-db-create-kdsms\" (UID: \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\") " pod="openstack/glance-db-create-kdsms" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.387959 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8ll4\" (UniqueName: \"kubernetes.io/projected/a643eeaf-1782-49f9-bbb7-55bdd019f68e-kube-api-access-r8ll4\") pod \"glance-db-create-kdsms\" (UID: \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\") " pod="openstack/glance-db-create-kdsms" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.398917 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-60b6-account-create-update-xwbqp"] Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.489758 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xsml\" (UniqueName: \"kubernetes.io/projected/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-kube-api-access-8xsml\") pod \"glance-60b6-account-create-update-xwbqp\" (UID: \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\") " pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.489863 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-operator-scripts\") pod \"glance-60b6-account-create-update-xwbqp\" (UID: \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\") " pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.489903 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a643eeaf-1782-49f9-bbb7-55bdd019f68e-operator-scripts\") pod \"glance-db-create-kdsms\" (UID: \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\") " pod="openstack/glance-db-create-kdsms" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.489945 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8ll4\" (UniqueName: \"kubernetes.io/projected/a643eeaf-1782-49f9-bbb7-55bdd019f68e-kube-api-access-r8ll4\") pod \"glance-db-create-kdsms\" (UID: \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\") " pod="openstack/glance-db-create-kdsms" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.490791 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a643eeaf-1782-49f9-bbb7-55bdd019f68e-operator-scripts\") pod \"glance-db-create-kdsms\" (UID: \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\") " pod="openstack/glance-db-create-kdsms" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.507488 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8ll4\" (UniqueName: \"kubernetes.io/projected/a643eeaf-1782-49f9-bbb7-55bdd019f68e-kube-api-access-r8ll4\") pod \"glance-db-create-kdsms\" (UID: \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\") " pod="openstack/glance-db-create-kdsms" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.591627 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-operator-scripts\") pod \"glance-60b6-account-create-update-xwbqp\" (UID: \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\") " pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.591772 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xsml\" (UniqueName: \"kubernetes.io/projected/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-kube-api-access-8xsml\") pod \"glance-60b6-account-create-update-xwbqp\" (UID: \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\") " pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.592782 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-operator-scripts\") pod \"glance-60b6-account-create-update-xwbqp\" (UID: \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\") " pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.612371 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kdsms" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.640510 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xsml\" (UniqueName: \"kubernetes.io/projected/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-kube-api-access-8xsml\") pod \"glance-60b6-account-create-update-xwbqp\" (UID: \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\") " pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.702413 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.706874 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d6139e2-68a5-4097-b8c2-cf11d11ed241" path="/var/lib/kubelet/pods/8d6139e2-68a5-4097-b8c2-cf11d11ed241/volumes" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.707947 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8520b34-412a-4d1c-a610-4c2e54df94bb" path="/var/lib/kubelet/pods/c8520b34-412a-4d1c-a610-4c2e54df94bb/volumes" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.750550 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.804418 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d909b510-1424-40ce-ad72-0b46a5304b42-operator-scripts\") pod \"d909b510-1424-40ce-ad72-0b46a5304b42\" (UID: \"d909b510-1424-40ce-ad72-0b46a5304b42\") " Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.804572 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq7kh\" (UniqueName: \"kubernetes.io/projected/d909b510-1424-40ce-ad72-0b46a5304b42-kube-api-access-kq7kh\") pod \"d909b510-1424-40ce-ad72-0b46a5304b42\" (UID: \"d909b510-1424-40ce-ad72-0b46a5304b42\") " Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.805104 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d909b510-1424-40ce-ad72-0b46a5304b42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d909b510-1424-40ce-ad72-0b46a5304b42" (UID: "d909b510-1424-40ce-ad72-0b46a5304b42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.805675 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d909b510-1424-40ce-ad72-0b46a5304b42-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.809825 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d909b510-1424-40ce-ad72-0b46a5304b42-kube-api-access-kq7kh" (OuterVolumeSpecName: "kube-api-access-kq7kh") pod "d909b510-1424-40ce-ad72-0b46a5304b42" (UID: "d909b510-1424-40ce-ad72-0b46a5304b42"). InnerVolumeSpecName "kube-api-access-kq7kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.909056 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq7kh\" (UniqueName: \"kubernetes.io/projected/d909b510-1424-40ce-ad72-0b46a5304b42-kube-api-access-kq7kh\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.936686 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.954708 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:09 crc kubenswrapper[4756]: I0203 09:27:09.982801 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.011859 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99vh4\" (UniqueName: \"kubernetes.io/projected/dc25ed95-beda-41da-b65f-10dea4ee003f-kube-api-access-99vh4\") pod \"dc25ed95-beda-41da-b65f-10dea4ee003f\" (UID: \"dc25ed95-beda-41da-b65f-10dea4ee003f\") " Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.012047 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-operator-scripts\") pod \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\" (UID: \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\") " Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.012088 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc25ed95-beda-41da-b65f-10dea4ee003f-operator-scripts\") pod \"dc25ed95-beda-41da-b65f-10dea4ee003f\" (UID: \"dc25ed95-beda-41da-b65f-10dea4ee003f\") " Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.012139 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r446\" (UniqueName: \"kubernetes.io/projected/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-kube-api-access-6r446\") pod \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\" (UID: \"4d05d9f3-e465-481b-92e0-c4d2c8c640bb\") " Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.012573 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d05d9f3-e465-481b-92e0-c4d2c8c640bb" (UID: "4d05d9f3-e465-481b-92e0-c4d2c8c640bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.012927 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc25ed95-beda-41da-b65f-10dea4ee003f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc25ed95-beda-41da-b65f-10dea4ee003f" (UID: "dc25ed95-beda-41da-b65f-10dea4ee003f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.016521 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc25ed95-beda-41da-b65f-10dea4ee003f-kube-api-access-99vh4" (OuterVolumeSpecName: "kube-api-access-99vh4") pod "dc25ed95-beda-41da-b65f-10dea4ee003f" (UID: "dc25ed95-beda-41da-b65f-10dea4ee003f"). InnerVolumeSpecName "kube-api-access-99vh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.017072 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-kube-api-access-6r446" (OuterVolumeSpecName: "kube-api-access-6r446") pod "4d05d9f3-e465-481b-92e0-c4d2c8c640bb" (UID: "4d05d9f3-e465-481b-92e0-c4d2c8c640bb"). InnerVolumeSpecName "kube-api-access-6r446". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.113077 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344e964e-c1ec-4fdc-82a3-223389b11f1a-operator-scripts\") pod \"344e964e-c1ec-4fdc-82a3-223389b11f1a\" (UID: \"344e964e-c1ec-4fdc-82a3-223389b11f1a\") " Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.113189 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl9jf\" (UniqueName: \"kubernetes.io/projected/344e964e-c1ec-4fdc-82a3-223389b11f1a-kube-api-access-kl9jf\") pod \"344e964e-c1ec-4fdc-82a3-223389b11f1a\" (UID: \"344e964e-c1ec-4fdc-82a3-223389b11f1a\") " Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.113626 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99vh4\" (UniqueName: \"kubernetes.io/projected/dc25ed95-beda-41da-b65f-10dea4ee003f-kube-api-access-99vh4\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.113640 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.113649 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc25ed95-beda-41da-b65f-10dea4ee003f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.113659 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r446\" (UniqueName: \"kubernetes.io/projected/4d05d9f3-e465-481b-92e0-c4d2c8c640bb-kube-api-access-6r446\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.114213 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/344e964e-c1ec-4fdc-82a3-223389b11f1a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "344e964e-c1ec-4fdc-82a3-223389b11f1a" (UID: "344e964e-c1ec-4fdc-82a3-223389b11f1a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.116410 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/344e964e-c1ec-4fdc-82a3-223389b11f1a-kube-api-access-kl9jf" (OuterVolumeSpecName: "kube-api-access-kl9jf") pod "344e964e-c1ec-4fdc-82a3-223389b11f1a" (UID: "344e964e-c1ec-4fdc-82a3-223389b11f1a"). InnerVolumeSpecName "kube-api-access-kl9jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.215420 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl9jf\" (UniqueName: \"kubernetes.io/projected/344e964e-c1ec-4fdc-82a3-223389b11f1a-kube-api-access-kl9jf\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.215542 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/344e964e-c1ec-4fdc-82a3-223389b11f1a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.241329 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kdsms"] Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.302841 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-60b6-account-create-update-xwbqp"] Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.303833 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-187d-account-create-update-8rkgm" event={"ID":"dc25ed95-beda-41da-b65f-10dea4ee003f","Type":"ContainerDied","Data":"1a85f6aedb1472a9fcd37280e66a3d1e1a4e5d21cde07b41a8b9e6c82872cc9d"} Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.303885 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a85f6aedb1472a9fcd37280e66a3d1e1a4e5d21cde07b41a8b9e6c82872cc9d" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.303937 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-187d-account-create-update-8rkgm" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.308196 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lzqmn" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.308252 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lzqmn" event={"ID":"344e964e-c1ec-4fdc-82a3-223389b11f1a","Type":"ContainerDied","Data":"8ec5597378e0d0c1d0dc501db8a5cb9c34db79e7d9ee6ce6ace3f0700209f013"} Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.308276 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ec5597378e0d0c1d0dc501db8a5cb9c34db79e7d9ee6ce6ace3f0700209f013" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.311332 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sl9vg" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.311368 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sl9vg" event={"ID":"d909b510-1424-40ce-ad72-0b46a5304b42","Type":"ContainerDied","Data":"3a710add519df44493a6a7c400e1b208108fa11bc2fc6240e9e83a8806c3f99d"} Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.311388 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a710add519df44493a6a7c400e1b208108fa11bc2fc6240e9e83a8806c3f99d" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.313248 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-aaef-account-create-update-cjsjc" event={"ID":"4d05d9f3-e465-481b-92e0-c4d2c8c640bb","Type":"ContainerDied","Data":"650eeea123d9a59a46e3dcd5c7cbca1a4b7f616a3c40d3cf5ab25ee7ba9f2807"} Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.313287 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="650eeea123d9a59a46e3dcd5c7cbca1a4b7f616a3c40d3cf5ab25ee7ba9f2807" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.313290 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-aaef-account-create-update-cjsjc" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.960978 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-7r875"] Feb 03 09:27:10 crc kubenswrapper[4756]: E0203 09:27:10.961978 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d909b510-1424-40ce-ad72-0b46a5304b42" containerName="mariadb-database-create" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962006 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d909b510-1424-40ce-ad72-0b46a5304b42" containerName="mariadb-database-create" Feb 03 09:27:10 crc kubenswrapper[4756]: E0203 09:27:10.962045 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="344e964e-c1ec-4fdc-82a3-223389b11f1a" containerName="mariadb-database-create" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962053 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="344e964e-c1ec-4fdc-82a3-223389b11f1a" containerName="mariadb-database-create" Feb 03 09:27:10 crc kubenswrapper[4756]: E0203 09:27:10.962087 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d05d9f3-e465-481b-92e0-c4d2c8c640bb" containerName="mariadb-account-create-update" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962097 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d05d9f3-e465-481b-92e0-c4d2c8c640bb" containerName="mariadb-account-create-update" Feb 03 09:27:10 crc kubenswrapper[4756]: E0203 09:27:10.962111 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc25ed95-beda-41da-b65f-10dea4ee003f" containerName="mariadb-account-create-update" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962118 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc25ed95-beda-41da-b65f-10dea4ee003f" containerName="mariadb-account-create-update" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962318 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d909b510-1424-40ce-ad72-0b46a5304b42" containerName="mariadb-database-create" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962339 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="344e964e-c1ec-4fdc-82a3-223389b11f1a" containerName="mariadb-database-create" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962355 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc25ed95-beda-41da-b65f-10dea4ee003f" containerName="mariadb-account-create-update" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962387 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d05d9f3-e465-481b-92e0-c4d2c8c640bb" containerName="mariadb-account-create-update" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.962954 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7r875" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.967858 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 03 09:27:10 crc kubenswrapper[4756]: I0203 09:27:10.980236 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-7r875"] Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.027465 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsb8v\" (UniqueName: \"kubernetes.io/projected/b24c11fe-462b-43e1-aebb-bf8639b4fd08-kube-api-access-fsb8v\") pod \"root-account-create-update-7r875\" (UID: \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\") " pod="openstack/root-account-create-update-7r875" Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.027570 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c11fe-462b-43e1-aebb-bf8639b4fd08-operator-scripts\") pod \"root-account-create-update-7r875\" (UID: \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\") " pod="openstack/root-account-create-update-7r875" Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.129234 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c11fe-462b-43e1-aebb-bf8639b4fd08-operator-scripts\") pod \"root-account-create-update-7r875\" (UID: \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\") " pod="openstack/root-account-create-update-7r875" Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.129347 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.129387 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsb8v\" (UniqueName: \"kubernetes.io/projected/b24c11fe-462b-43e1-aebb-bf8639b4fd08-kube-api-access-fsb8v\") pod \"root-account-create-update-7r875\" (UID: \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\") " pod="openstack/root-account-create-update-7r875" Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.130503 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c11fe-462b-43e1-aebb-bf8639b4fd08-operator-scripts\") pod \"root-account-create-update-7r875\" (UID: \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\") " pod="openstack/root-account-create-update-7r875" Feb 03 09:27:11 crc kubenswrapper[4756]: E0203 09:27:11.130618 4756 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 03 09:27:11 crc kubenswrapper[4756]: E0203 09:27:11.130638 4756 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 03 09:27:11 crc kubenswrapper[4756]: E0203 09:27:11.130677 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift podName:d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9 nodeName:}" failed. No retries permitted until 2026-02-03 09:27:15.130663598 +0000 UTC m=+1026.281130973 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift") pod "swift-storage-0" (UID: "d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9") : configmap "swift-ring-files" not found Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.149534 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsb8v\" (UniqueName: \"kubernetes.io/projected/b24c11fe-462b-43e1-aebb-bf8639b4fd08-kube-api-access-fsb8v\") pod \"root-account-create-update-7r875\" (UID: \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\") " pod="openstack/root-account-create-update-7r875" Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.299294 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7r875" Feb 03 09:27:11 crc kubenswrapper[4756]: W0203 09:27:11.798551 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0729c9f_8d1d_49a8_9cff_eefbcafac7b0.slice/crio-d381bd30b8763687d8bf51b91e9edbefdcaabb8b3308cbb1b8f3558e2ea2558b WatchSource:0}: Error finding container d381bd30b8763687d8bf51b91e9edbefdcaabb8b3308cbb1b8f3558e2ea2558b: Status 404 returned error can't find the container with id d381bd30b8763687d8bf51b91e9edbefdcaabb8b3308cbb1b8f3558e2ea2558b Feb 03 09:27:11 crc kubenswrapper[4756]: W0203 09:27:11.801525 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda643eeaf_1782_49f9_bbb7_55bdd019f68e.slice/crio-dcbe5df344b12632e1d7e687d4e7555a7b9ec183d5f07b45ccca98d67169a74e WatchSource:0}: Error finding container dcbe5df344b12632e1d7e687d4e7555a7b9ec183d5f07b45ccca98d67169a74e: Status 404 returned error can't find the container with id dcbe5df344b12632e1d7e687d4e7555a7b9ec183d5f07b45ccca98d67169a74e Feb 03 09:27:11 crc kubenswrapper[4756]: I0203 09:27:11.869068 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.267902 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-7r875"] Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.332012 4756 generic.go:334] "Generic (PLEG): container finished" podID="a643eeaf-1782-49f9-bbb7-55bdd019f68e" containerID="e94495bb22c6c2245ab14c6e8c70298201b7820743e32866bbf019e7ba9f0112" exitCode=0 Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.332059 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kdsms" event={"ID":"a643eeaf-1782-49f9-bbb7-55bdd019f68e","Type":"ContainerDied","Data":"e94495bb22c6c2245ab14c6e8c70298201b7820743e32866bbf019e7ba9f0112"} Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.332106 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kdsms" event={"ID":"a643eeaf-1782-49f9-bbb7-55bdd019f68e","Type":"ContainerStarted","Data":"dcbe5df344b12632e1d7e687d4e7555a7b9ec183d5f07b45ccca98d67169a74e"} Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.334137 4756 generic.go:334] "Generic (PLEG): container finished" podID="f0729c9f-8d1d-49a8-9cff-eefbcafac7b0" containerID="de0dcf48427ffdf38b3cb5ad420617b2892e5f15c91ff1d65746372eff9f8f41" exitCode=0 Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.334183 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-60b6-account-create-update-xwbqp" event={"ID":"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0","Type":"ContainerDied","Data":"de0dcf48427ffdf38b3cb5ad420617b2892e5f15c91ff1d65746372eff9f8f41"} Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.334249 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-60b6-account-create-update-xwbqp" event={"ID":"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0","Type":"ContainerStarted","Data":"d381bd30b8763687d8bf51b91e9edbefdcaabb8b3308cbb1b8f3558e2ea2558b"} Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.337861 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pqfmb" event={"ID":"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e","Type":"ContainerStarted","Data":"141d8ec967d1d88637ce4ca21a9c3a9e113985b0af654c1297f53e400183fc49"} Feb 03 09:27:12 crc kubenswrapper[4756]: I0203 09:27:12.384852 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-pqfmb" podStartSLOduration=1.859616704 podStartE2EDuration="5.384825654s" podCreationTimestamp="2026-02-03 09:27:07 +0000 UTC" firstStartedPulling="2026-02-03 09:27:08.367398836 +0000 UTC m=+1019.517866211" lastFinishedPulling="2026-02-03 09:27:11.892607786 +0000 UTC m=+1023.043075161" observedRunningTime="2026-02-03 09:27:12.366823502 +0000 UTC m=+1023.517290877" watchObservedRunningTime="2026-02-03 09:27:12.384825654 +0000 UTC m=+1023.535293069" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.346513 4756 generic.go:334] "Generic (PLEG): container finished" podID="b24c11fe-462b-43e1-aebb-bf8639b4fd08" containerID="3fae684069ea3d1847d0fb5eff949e9e79247538bc2dd886e0c6bf6ba83cb4a4" exitCode=0 Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.346559 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7r875" event={"ID":"b24c11fe-462b-43e1-aebb-bf8639b4fd08","Type":"ContainerDied","Data":"3fae684069ea3d1847d0fb5eff949e9e79247538bc2dd886e0c6bf6ba83cb4a4"} Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.346845 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7r875" event={"ID":"b24c11fe-462b-43e1-aebb-bf8639b4fd08","Type":"ContainerStarted","Data":"7d8a4617e14ada54fade7db5637507bb8d9c1d84f0bd99a6f6f0c4a91478f064"} Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.748247 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.754933 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kdsms" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.808588 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.876887 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a643eeaf-1782-49f9-bbb7-55bdd019f68e-operator-scripts\") pod \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\" (UID: \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\") " Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.877047 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-operator-scripts\") pod \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\" (UID: \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\") " Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.877094 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8ll4\" (UniqueName: \"kubernetes.io/projected/a643eeaf-1782-49f9-bbb7-55bdd019f68e-kube-api-access-r8ll4\") pod \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\" (UID: \"a643eeaf-1782-49f9-bbb7-55bdd019f68e\") " Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.877178 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xsml\" (UniqueName: \"kubernetes.io/projected/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-kube-api-access-8xsml\") pod \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\" (UID: \"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0\") " Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.877884 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0729c9f-8d1d-49a8-9cff-eefbcafac7b0" (UID: "f0729c9f-8d1d-49a8-9cff-eefbcafac7b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.877905 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a643eeaf-1782-49f9-bbb7-55bdd019f68e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a643eeaf-1782-49f9-bbb7-55bdd019f68e" (UID: "a643eeaf-1782-49f9-bbb7-55bdd019f68e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.883954 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-kube-api-access-8xsml" (OuterVolumeSpecName: "kube-api-access-8xsml") pod "f0729c9f-8d1d-49a8-9cff-eefbcafac7b0" (UID: "f0729c9f-8d1d-49a8-9cff-eefbcafac7b0"). InnerVolumeSpecName "kube-api-access-8xsml". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.884097 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a643eeaf-1782-49f9-bbb7-55bdd019f68e-kube-api-access-r8ll4" (OuterVolumeSpecName: "kube-api-access-r8ll4") pod "a643eeaf-1782-49f9-bbb7-55bdd019f68e" (UID: "a643eeaf-1782-49f9-bbb7-55bdd019f68e"). InnerVolumeSpecName "kube-api-access-r8ll4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.979563 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a643eeaf-1782-49f9-bbb7-55bdd019f68e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.979595 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.979604 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8ll4\" (UniqueName: \"kubernetes.io/projected/a643eeaf-1782-49f9-bbb7-55bdd019f68e-kube-api-access-r8ll4\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:13 crc kubenswrapper[4756]: I0203 09:27:13.979615 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xsml\" (UniqueName: \"kubernetes.io/projected/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0-kube-api-access-8xsml\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.359349 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-60b6-account-create-update-xwbqp" event={"ID":"f0729c9f-8d1d-49a8-9cff-eefbcafac7b0","Type":"ContainerDied","Data":"d381bd30b8763687d8bf51b91e9edbefdcaabb8b3308cbb1b8f3558e2ea2558b"} Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.361061 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d381bd30b8763687d8bf51b91e9edbefdcaabb8b3308cbb1b8f3558e2ea2558b" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.360197 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-60b6-account-create-update-xwbqp" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.362394 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kdsms" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.362396 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kdsms" event={"ID":"a643eeaf-1782-49f9-bbb7-55bdd019f68e","Type":"ContainerDied","Data":"dcbe5df344b12632e1d7e687d4e7555a7b9ec183d5f07b45ccca98d67169a74e"} Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.362662 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcbe5df344b12632e1d7e687d4e7555a7b9ec183d5f07b45ccca98d67169a74e" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.719415 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7r875" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.789787 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c11fe-462b-43e1-aebb-bf8639b4fd08-operator-scripts\") pod \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\" (UID: \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\") " Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.789900 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsb8v\" (UniqueName: \"kubernetes.io/projected/b24c11fe-462b-43e1-aebb-bf8639b4fd08-kube-api-access-fsb8v\") pod \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\" (UID: \"b24c11fe-462b-43e1-aebb-bf8639b4fd08\") " Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.790238 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b24c11fe-462b-43e1-aebb-bf8639b4fd08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b24c11fe-462b-43e1-aebb-bf8639b4fd08" (UID: "b24c11fe-462b-43e1-aebb-bf8639b4fd08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.790411 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c11fe-462b-43e1-aebb-bf8639b4fd08-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.795654 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24c11fe-462b-43e1-aebb-bf8639b4fd08-kube-api-access-fsb8v" (OuterVolumeSpecName: "kube-api-access-fsb8v") pod "b24c11fe-462b-43e1-aebb-bf8639b4fd08" (UID: "b24c11fe-462b-43e1-aebb-bf8639b4fd08"). InnerVolumeSpecName "kube-api-access-fsb8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:14 crc kubenswrapper[4756]: I0203 09:27:14.891830 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsb8v\" (UniqueName: \"kubernetes.io/projected/b24c11fe-462b-43e1-aebb-bf8639b4fd08-kube-api-access-fsb8v\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:15 crc kubenswrapper[4756]: I0203 09:27:15.195184 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:15 crc kubenswrapper[4756]: E0203 09:27:15.195378 4756 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 03 09:27:15 crc kubenswrapper[4756]: E0203 09:27:15.195416 4756 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 03 09:27:15 crc kubenswrapper[4756]: E0203 09:27:15.195501 4756 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift podName:d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9 nodeName:}" failed. No retries permitted until 2026-02-03 09:27:23.195476083 +0000 UTC m=+1034.345943458 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift") pod "swift-storage-0" (UID: "d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9") : configmap "swift-ring-files" not found Feb 03 09:27:15 crc kubenswrapper[4756]: I0203 09:27:15.371775 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7r875" event={"ID":"b24c11fe-462b-43e1-aebb-bf8639b4fd08","Type":"ContainerDied","Data":"7d8a4617e14ada54fade7db5637507bb8d9c1d84f0bd99a6f6f0c4a91478f064"} Feb 03 09:27:15 crc kubenswrapper[4756]: I0203 09:27:15.371814 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d8a4617e14ada54fade7db5637507bb8d9c1d84f0bd99a6f6f0c4a91478f064" Feb 03 09:27:15 crc kubenswrapper[4756]: I0203 09:27:15.371846 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7r875" Feb 03 09:27:16 crc kubenswrapper[4756]: I0203 09:27:16.469219 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:16 crc kubenswrapper[4756]: I0203 09:27:16.537036 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwr6s"] Feb 03 09:27:16 crc kubenswrapper[4756]: I0203 09:27:16.537320 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" podUID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" containerName="dnsmasq-dns" containerID="cri-o://2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48" gracePeriod=10 Feb 03 09:27:16 crc kubenswrapper[4756]: I0203 09:27:16.988531 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.078548 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.131004 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqpwk\" (UniqueName: \"kubernetes.io/projected/d1dbab7f-1fdd-49f6-a7a6-904feb481959-kube-api-access-cqpwk\") pod \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.131114 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-config\") pod \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.131133 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-dns-svc\") pod \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\" (UID: \"d1dbab7f-1fdd-49f6-a7a6-904feb481959\") " Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.137336 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1dbab7f-1fdd-49f6-a7a6-904feb481959-kube-api-access-cqpwk" (OuterVolumeSpecName: "kube-api-access-cqpwk") pod "d1dbab7f-1fdd-49f6-a7a6-904feb481959" (UID: "d1dbab7f-1fdd-49f6-a7a6-904feb481959"). InnerVolumeSpecName "kube-api-access-cqpwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.177193 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-config" (OuterVolumeSpecName: "config") pod "d1dbab7f-1fdd-49f6-a7a6-904feb481959" (UID: "d1dbab7f-1fdd-49f6-a7a6-904feb481959"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.205113 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d1dbab7f-1fdd-49f6-a7a6-904feb481959" (UID: "d1dbab7f-1fdd-49f6-a7a6-904feb481959"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.232730 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.232768 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1dbab7f-1fdd-49f6-a7a6-904feb481959-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.232781 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqpwk\" (UniqueName: \"kubernetes.io/projected/d1dbab7f-1fdd-49f6-a7a6-904feb481959-kube-api-access-cqpwk\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.388655 4756 generic.go:334] "Generic (PLEG): container finished" podID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" containerID="2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48" exitCode=0 Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.388710 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.388709 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" event={"ID":"d1dbab7f-1fdd-49f6-a7a6-904feb481959","Type":"ContainerDied","Data":"2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48"} Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.388853 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwr6s" event={"ID":"d1dbab7f-1fdd-49f6-a7a6-904feb481959","Type":"ContainerDied","Data":"8d34da076b63a07b58e3b4aa772da20d4c1dc54d4f25db8d8d5e38dc8b2beeee"} Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.388884 4756 scope.go:117] "RemoveContainer" containerID="2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.408101 4756 scope.go:117] "RemoveContainer" containerID="de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.430464 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwr6s"] Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.437809 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwr6s"] Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.441040 4756 scope.go:117] "RemoveContainer" containerID="2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48" Feb 03 09:27:17 crc kubenswrapper[4756]: E0203 09:27:17.441495 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48\": container with ID starting with 2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48 not found: ID does not exist" containerID="2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.441557 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48"} err="failed to get container status \"2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48\": rpc error: code = NotFound desc = could not find container \"2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48\": container with ID starting with 2a4d0ed27f30decbfa8861b64517a6476d86ccf3d894c16167e66431f2142a48 not found: ID does not exist" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.441590 4756 scope.go:117] "RemoveContainer" containerID="de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0" Feb 03 09:27:17 crc kubenswrapper[4756]: E0203 09:27:17.441934 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0\": container with ID starting with de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0 not found: ID does not exist" containerID="de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.441977 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0"} err="failed to get container status \"de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0\": rpc error: code = NotFound desc = could not find container \"de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0\": container with ID starting with de2a4a99c3331866bb2f0a1d6642ca149f4f147b73900491f839d1b600350fa0 not found: ID does not exist" Feb 03 09:27:17 crc kubenswrapper[4756]: I0203 09:27:17.632120 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" path="/var/lib/kubelet/pods/d1dbab7f-1fdd-49f6-a7a6-904feb481959/volumes" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.402974 4756 generic.go:334] "Generic (PLEG): container finished" podID="f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" containerID="141d8ec967d1d88637ce4ca21a9c3a9e113985b0af654c1297f53e400183fc49" exitCode=0 Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.403047 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pqfmb" event={"ID":"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e","Type":"ContainerDied","Data":"141d8ec967d1d88637ce4ca21a9c3a9e113985b0af654c1297f53e400183fc49"} Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623236 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-8j6l5"] Feb 03 09:27:19 crc kubenswrapper[4756]: E0203 09:27:19.623514 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0729c9f-8d1d-49a8-9cff-eefbcafac7b0" containerName="mariadb-account-create-update" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623532 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0729c9f-8d1d-49a8-9cff-eefbcafac7b0" containerName="mariadb-account-create-update" Feb 03 09:27:19 crc kubenswrapper[4756]: E0203 09:27:19.623544 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a643eeaf-1782-49f9-bbb7-55bdd019f68e" containerName="mariadb-database-create" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623551 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a643eeaf-1782-49f9-bbb7-55bdd019f68e" containerName="mariadb-database-create" Feb 03 09:27:19 crc kubenswrapper[4756]: E0203 09:27:19.623563 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" containerName="dnsmasq-dns" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623570 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" containerName="dnsmasq-dns" Feb 03 09:27:19 crc kubenswrapper[4756]: E0203 09:27:19.623589 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24c11fe-462b-43e1-aebb-bf8639b4fd08" containerName="mariadb-account-create-update" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623594 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24c11fe-462b-43e1-aebb-bf8639b4fd08" containerName="mariadb-account-create-update" Feb 03 09:27:19 crc kubenswrapper[4756]: E0203 09:27:19.623605 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" containerName="init" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623611 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" containerName="init" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623770 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a643eeaf-1782-49f9-bbb7-55bdd019f68e" containerName="mariadb-database-create" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623790 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0729c9f-8d1d-49a8-9cff-eefbcafac7b0" containerName="mariadb-account-create-update" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623808 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24c11fe-462b-43e1-aebb-bf8639b4fd08" containerName="mariadb-account-create-update" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.623819 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1dbab7f-1fdd-49f6-a7a6-904feb481959" containerName="dnsmasq-dns" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.624313 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.629840 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-78t56" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.629965 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.635586 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8j6l5"] Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.669257 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-combined-ca-bundle\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.669295 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-db-sync-config-data\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.669357 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq7nr\" (UniqueName: \"kubernetes.io/projected/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-kube-api-access-kq7nr\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.669374 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-config-data\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.770875 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-combined-ca-bundle\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.770960 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-db-sync-config-data\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.771023 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq7nr\" (UniqueName: \"kubernetes.io/projected/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-kube-api-access-kq7nr\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.771047 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-config-data\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.777161 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-db-sync-config-data\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.777245 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-config-data\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.778572 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-combined-ca-bundle\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.789058 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq7nr\" (UniqueName: \"kubernetes.io/projected/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-kube-api-access-kq7nr\") pod \"glance-db-sync-8j6l5\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:19 crc kubenswrapper[4756]: I0203 09:27:19.947659 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.454490 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8j6l5"] Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.677051 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.784323 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-etc-swift\") pod \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.784466 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-dispersionconf\") pod \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.784549 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-scripts\") pod \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.784584 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l52z\" (UniqueName: \"kubernetes.io/projected/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-kube-api-access-7l52z\") pod \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.784614 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-swiftconf\") pod \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.784640 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-combined-ca-bundle\") pod \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.784685 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-ring-data-devices\") pod \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\" (UID: \"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e\") " Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.786094 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" (UID: "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.786857 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" (UID: "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.794578 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-kube-api-access-7l52z" (OuterVolumeSpecName: "kube-api-access-7l52z") pod "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" (UID: "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e"). InnerVolumeSpecName "kube-api-access-7l52z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.794946 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" (UID: "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.816390 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-scripts" (OuterVolumeSpecName: "scripts") pod "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" (UID: "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.816682 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" (UID: "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.817110 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" (UID: "f356b5c0-fae0-4a67-9624-4fbe59bd2f9e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.886214 4756 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.886247 4756 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.886259 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.886268 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l52z\" (UniqueName: \"kubernetes.io/projected/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-kube-api-access-7l52z\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.886276 4756 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.886284 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:20 crc kubenswrapper[4756]: I0203 09:27:20.886292 4756 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f356b5c0-fae0-4a67-9624-4fbe59bd2f9e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:21 crc kubenswrapper[4756]: I0203 09:27:21.416315 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8j6l5" event={"ID":"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9","Type":"ContainerStarted","Data":"33938a62e18a373c2dae7174343ee1d4757cb94fb90d209e0acd3f66f5ccb7af"} Feb 03 09:27:21 crc kubenswrapper[4756]: I0203 09:27:21.417536 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pqfmb" event={"ID":"f356b5c0-fae0-4a67-9624-4fbe59bd2f9e","Type":"ContainerDied","Data":"56bab4967dbd539cf2aa6bcc439d6eb2b6595bdc07ceb312bd2fee6e6ccd131c"} Feb 03 09:27:21 crc kubenswrapper[4756]: I0203 09:27:21.417559 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56bab4967dbd539cf2aa6bcc439d6eb2b6595bdc07ceb312bd2fee6e6ccd131c" Feb 03 09:27:21 crc kubenswrapper[4756]: I0203 09:27:21.417605 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pqfmb" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.365635 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-7r875"] Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.373575 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-7r875"] Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.440227 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-bcd5t"] Feb 03 09:27:22 crc kubenswrapper[4756]: E0203 09:27:22.440684 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" containerName="swift-ring-rebalance" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.440707 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" containerName="swift-ring-rebalance" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.440912 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f356b5c0-fae0-4a67-9624-4fbe59bd2f9e" containerName="swift-ring-rebalance" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.441787 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.449947 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.451918 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bcd5t"] Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.516553 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx2kp\" (UniqueName: \"kubernetes.io/projected/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-kube-api-access-dx2kp\") pod \"root-account-create-update-bcd5t\" (UID: \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\") " pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.516845 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-operator-scripts\") pod \"root-account-create-update-bcd5t\" (UID: \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\") " pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.618165 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx2kp\" (UniqueName: \"kubernetes.io/projected/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-kube-api-access-dx2kp\") pod \"root-account-create-update-bcd5t\" (UID: \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\") " pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.618259 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-operator-scripts\") pod \"root-account-create-update-bcd5t\" (UID: \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\") " pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.619723 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-operator-scripts\") pod \"root-account-create-update-bcd5t\" (UID: \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\") " pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.640238 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx2kp\" (UniqueName: \"kubernetes.io/projected/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-kube-api-access-dx2kp\") pod \"root-account-create-update-bcd5t\" (UID: \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\") " pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:22 crc kubenswrapper[4756]: I0203 09:27:22.778522 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.158015 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bcd5t"] Feb 03 09:27:23 crc kubenswrapper[4756]: W0203 09:27:23.172887 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa4bceb9_3d5a_4fe4_9b98_ed408d36116b.slice/crio-38af79613236429b8714fee1563611f831508b6b343680efe137d261c48355ba WatchSource:0}: Error finding container 38af79613236429b8714fee1563611f831508b6b343680efe137d261c48355ba: Status 404 returned error can't find the container with id 38af79613236429b8714fee1563611f831508b6b343680efe137d261c48355ba Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.233882 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.243089 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9-etc-swift\") pod \"swift-storage-0\" (UID: \"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9\") " pod="openstack/swift-storage-0" Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.274670 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.435344 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bcd5t" event={"ID":"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b","Type":"ContainerStarted","Data":"891ceadb7fb03f03a5e991cb3032159128fc9153668570a7cf316d803966f9e1"} Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.435393 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bcd5t" event={"ID":"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b","Type":"ContainerStarted","Data":"38af79613236429b8714fee1563611f831508b6b343680efe137d261c48355ba"} Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.472907 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-bcd5t" podStartSLOduration=1.472888176 podStartE2EDuration="1.472888176s" podCreationTimestamp="2026-02-03 09:27:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:23.466330248 +0000 UTC m=+1034.616797623" watchObservedRunningTime="2026-02-03 09:27:23.472888176 +0000 UTC m=+1034.623355551" Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.631884 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24c11fe-462b-43e1-aebb-bf8639b4fd08" path="/var/lib/kubelet/pods/b24c11fe-462b-43e1-aebb-bf8639b4fd08/volumes" Feb 03 09:27:23 crc kubenswrapper[4756]: I0203 09:27:23.867010 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.443782 4756 generic.go:334] "Generic (PLEG): container finished" podID="aa4bceb9-3d5a-4fe4-9b98-ed408d36116b" containerID="891ceadb7fb03f03a5e991cb3032159128fc9153668570a7cf316d803966f9e1" exitCode=0 Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.443852 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bcd5t" event={"ID":"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b","Type":"ContainerDied","Data":"891ceadb7fb03f03a5e991cb3032159128fc9153668570a7cf316d803966f9e1"} Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.445163 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"3d09643f44efd3f31f6917b6b6af666b1989ffdd928da9900afd3fc9fbad02ad"} Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.447845 4756 generic.go:334] "Generic (PLEG): container finished" podID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerID="ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f" exitCode=0 Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.447907 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c","Type":"ContainerDied","Data":"ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f"} Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.451905 4756 generic.go:334] "Generic (PLEG): container finished" podID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerID="30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53" exitCode=0 Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.451950 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f2aeafe0-b129-44d5-86d9-0a95a7ac180d","Type":"ContainerDied","Data":"30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53"} Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.741291 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-wnf95" podUID="15695b6e-11fa-4913-b502-a78974ddf81c" containerName="ovn-controller" probeResult="failure" output=< Feb 03 09:27:24 crc kubenswrapper[4756]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 03 09:27:24 crc kubenswrapper[4756]: > Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.760977 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:27:24 crc kubenswrapper[4756]: I0203 09:27:24.792871 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fb566" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.008111 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-wnf95-config-l5vxr"] Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.013954 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.016947 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.024093 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wnf95-config-l5vxr"] Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.168887 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.168978 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf29g\" (UniqueName: \"kubernetes.io/projected/6930d990-1321-4ec0-9422-91f1b5abbf5a-kube-api-access-rf29g\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.169035 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-scripts\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.169057 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run-ovn\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.169078 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-log-ovn\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.169145 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-additional-scripts\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.270544 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf29g\" (UniqueName: \"kubernetes.io/projected/6930d990-1321-4ec0-9422-91f1b5abbf5a-kube-api-access-rf29g\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.270684 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-scripts\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.270719 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run-ovn\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.270743 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-log-ovn\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.270784 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-additional-scripts\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.270818 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.270987 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run-ovn\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.271043 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.271069 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-log-ovn\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.272014 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-additional-scripts\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.272915 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-scripts\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.292044 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf29g\" (UniqueName: \"kubernetes.io/projected/6930d990-1321-4ec0-9422-91f1b5abbf5a-kube-api-access-rf29g\") pod \"ovn-controller-wnf95-config-l5vxr\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.334501 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.463635 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c","Type":"ContainerStarted","Data":"fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4"} Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.465131 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.474761 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f2aeafe0-b129-44d5-86d9-0a95a7ac180d","Type":"ContainerStarted","Data":"1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f"} Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.475668 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.551064 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=49.362942154 podStartE2EDuration="56.551045394s" podCreationTimestamp="2026-02-03 09:26:29 +0000 UTC" firstStartedPulling="2026-02-03 09:26:43.800226782 +0000 UTC m=+994.950694157" lastFinishedPulling="2026-02-03 09:26:50.988330022 +0000 UTC m=+1002.138797397" observedRunningTime="2026-02-03 09:27:25.542148292 +0000 UTC m=+1036.692615657" watchObservedRunningTime="2026-02-03 09:27:25.551045394 +0000 UTC m=+1036.701512769" Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.591785 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=49.292807563 podStartE2EDuration="56.59176874s" podCreationTimestamp="2026-02-03 09:26:29 +0000 UTC" firstStartedPulling="2026-02-03 09:26:43.740181332 +0000 UTC m=+994.890648707" lastFinishedPulling="2026-02-03 09:26:51.039142509 +0000 UTC m=+1002.189609884" observedRunningTime="2026-02-03 09:27:25.580877824 +0000 UTC m=+1036.731345199" watchObservedRunningTime="2026-02-03 09:27:25.59176874 +0000 UTC m=+1036.742236115" Feb 03 09:27:25 crc kubenswrapper[4756]: W0203 09:27:25.922698 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6930d990_1321_4ec0_9422_91f1b5abbf5a.slice/crio-b0bce03cca2f9abae1a6d7ac096480743ae38cbf0d74da459e5c2b0effb8cefc WatchSource:0}: Error finding container b0bce03cca2f9abae1a6d7ac096480743ae38cbf0d74da459e5c2b0effb8cefc: Status 404 returned error can't find the container with id b0bce03cca2f9abae1a6d7ac096480743ae38cbf0d74da459e5c2b0effb8cefc Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.927975 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wnf95-config-l5vxr"] Feb 03 09:27:25 crc kubenswrapper[4756]: I0203 09:27:25.976720 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.087771 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-operator-scripts\") pod \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\" (UID: \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\") " Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.088036 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx2kp\" (UniqueName: \"kubernetes.io/projected/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-kube-api-access-dx2kp\") pod \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\" (UID: \"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b\") " Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.088511 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aa4bceb9-3d5a-4fe4-9b98-ed408d36116b" (UID: "aa4bceb9-3d5a-4fe4-9b98-ed408d36116b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.089352 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.093588 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-kube-api-access-dx2kp" (OuterVolumeSpecName: "kube-api-access-dx2kp") pod "aa4bceb9-3d5a-4fe4-9b98-ed408d36116b" (UID: "aa4bceb9-3d5a-4fe4-9b98-ed408d36116b"). InnerVolumeSpecName "kube-api-access-dx2kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.191163 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx2kp\" (UniqueName: \"kubernetes.io/projected/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b-kube-api-access-dx2kp\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.497912 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"eb66b74e106a9f217d4987cd2996f44c05c15d029b8706038ba16ab47b5b98d8"} Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.498264 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"a9a0564a2ae45f2cca828374f9cef94e01c000937537e284646cb1bae6583a8a"} Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.498279 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"f37cf6cc00d9ed99699d7569327ebd4259427a68031c01718c455f63d888a55c"} Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.498291 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"3eb24e3c40722b0c878e182f0619a3e7f928c294741c7e34577f56897564dd64"} Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.505688 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95-config-l5vxr" event={"ID":"6930d990-1321-4ec0-9422-91f1b5abbf5a","Type":"ContainerStarted","Data":"9fccfad52bb4f0de77a6b4eb954b6a24385b5459d1df325471e70950abb9595d"} Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.505734 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95-config-l5vxr" event={"ID":"6930d990-1321-4ec0-9422-91f1b5abbf5a","Type":"ContainerStarted","Data":"b0bce03cca2f9abae1a6d7ac096480743ae38cbf0d74da459e5c2b0effb8cefc"} Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.507923 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bcd5t" Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.508212 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bcd5t" event={"ID":"aa4bceb9-3d5a-4fe4-9b98-ed408d36116b","Type":"ContainerDied","Data":"38af79613236429b8714fee1563611f831508b6b343680efe137d261c48355ba"} Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.508258 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38af79613236429b8714fee1563611f831508b6b343680efe137d261c48355ba" Feb 03 09:27:26 crc kubenswrapper[4756]: I0203 09:27:26.525055 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-wnf95-config-l5vxr" podStartSLOduration=2.525037517 podStartE2EDuration="2.525037517s" podCreationTimestamp="2026-02-03 09:27:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:26.522682563 +0000 UTC m=+1037.673149938" watchObservedRunningTime="2026-02-03 09:27:26.525037517 +0000 UTC m=+1037.675504902" Feb 03 09:27:27 crc kubenswrapper[4756]: I0203 09:27:27.521148 4756 generic.go:334] "Generic (PLEG): container finished" podID="6930d990-1321-4ec0-9422-91f1b5abbf5a" containerID="9fccfad52bb4f0de77a6b4eb954b6a24385b5459d1df325471e70950abb9595d" exitCode=0 Feb 03 09:27:27 crc kubenswrapper[4756]: I0203 09:27:27.521204 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95-config-l5vxr" event={"ID":"6930d990-1321-4ec0-9422-91f1b5abbf5a","Type":"ContainerDied","Data":"9fccfad52bb4f0de77a6b4eb954b6a24385b5459d1df325471e70950abb9595d"} Feb 03 09:27:29 crc kubenswrapper[4756]: I0203 09:27:29.717032 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-wnf95" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.495555 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.596611 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95-config-l5vxr" event={"ID":"6930d990-1321-4ec0-9422-91f1b5abbf5a","Type":"ContainerDied","Data":"b0bce03cca2f9abae1a6d7ac096480743ae38cbf0d74da459e5c2b0effb8cefc"} Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.596890 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0bce03cca2f9abae1a6d7ac096480743ae38cbf0d74da459e5c2b0effb8cefc" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.596824 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95-config-l5vxr" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.605581 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run-ovn\") pod \"6930d990-1321-4ec0-9422-91f1b5abbf5a\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.605630 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run\") pod \"6930d990-1321-4ec0-9422-91f1b5abbf5a\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.605658 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-scripts\") pod \"6930d990-1321-4ec0-9422-91f1b5abbf5a\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.605678 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-log-ovn\") pod \"6930d990-1321-4ec0-9422-91f1b5abbf5a\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.605683 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6930d990-1321-4ec0-9422-91f1b5abbf5a" (UID: "6930d990-1321-4ec0-9422-91f1b5abbf5a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.605702 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-additional-scripts\") pod \"6930d990-1321-4ec0-9422-91f1b5abbf5a\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.605749 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf29g\" (UniqueName: \"kubernetes.io/projected/6930d990-1321-4ec0-9422-91f1b5abbf5a-kube-api-access-rf29g\") pod \"6930d990-1321-4ec0-9422-91f1b5abbf5a\" (UID: \"6930d990-1321-4ec0-9422-91f1b5abbf5a\") " Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.606015 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run" (OuterVolumeSpecName: "var-run") pod "6930d990-1321-4ec0-9422-91f1b5abbf5a" (UID: "6930d990-1321-4ec0-9422-91f1b5abbf5a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.606389 4756 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.606409 4756 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-run\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.606438 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6930d990-1321-4ec0-9422-91f1b5abbf5a" (UID: "6930d990-1321-4ec0-9422-91f1b5abbf5a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.606847 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6930d990-1321-4ec0-9422-91f1b5abbf5a" (UID: "6930d990-1321-4ec0-9422-91f1b5abbf5a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.607137 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-scripts" (OuterVolumeSpecName: "scripts") pod "6930d990-1321-4ec0-9422-91f1b5abbf5a" (UID: "6930d990-1321-4ec0-9422-91f1b5abbf5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.609666 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6930d990-1321-4ec0-9422-91f1b5abbf5a-kube-api-access-rf29g" (OuterVolumeSpecName: "kube-api-access-rf29g") pod "6930d990-1321-4ec0-9422-91f1b5abbf5a" (UID: "6930d990-1321-4ec0-9422-91f1b5abbf5a"). InnerVolumeSpecName "kube-api-access-rf29g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.707365 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.707394 4756 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6930d990-1321-4ec0-9422-91f1b5abbf5a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.707406 4756 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6930d990-1321-4ec0-9422-91f1b5abbf5a-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:33 crc kubenswrapper[4756]: I0203 09:27:33.707419 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf29g\" (UniqueName: \"kubernetes.io/projected/6930d990-1321-4ec0-9422-91f1b5abbf5a-kube-api-access-rf29g\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.613510 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"8ba2d02b001eaa50a14bac1b76b2816cf9994436a92714f6f8f7bb009041a7c6"} Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.613817 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"7d12c87ccda293e8015c2d36e0c3b7992029d6591ce55752c3d0d26a870e8190"} Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.613828 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"fad9be8eabc19ee66a6508a15dcdb16192e7ebd758762eb9b592549b2cdc13a6"} Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.615064 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8j6l5" event={"ID":"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9","Type":"ContainerStarted","Data":"57832a58c9733cca830b6ae89f7f260772e41d1643546ecd029a943252c27533"} Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.635037 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-wnf95-config-l5vxr"] Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.641349 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-wnf95-config-l5vxr"] Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.655158 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-8j6l5" podStartSLOduration=2.607766215 podStartE2EDuration="15.655135804s" podCreationTimestamp="2026-02-03 09:27:19 +0000 UTC" firstStartedPulling="2026-02-03 09:27:20.464144145 +0000 UTC m=+1031.614611520" lastFinishedPulling="2026-02-03 09:27:33.511513734 +0000 UTC m=+1044.661981109" observedRunningTime="2026-02-03 09:27:34.633561258 +0000 UTC m=+1045.784028643" watchObservedRunningTime="2026-02-03 09:27:34.655135804 +0000 UTC m=+1045.805603179" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.812192 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-wnf95-config-p5bn8"] Feb 03 09:27:34 crc kubenswrapper[4756]: E0203 09:27:34.812536 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6930d990-1321-4ec0-9422-91f1b5abbf5a" containerName="ovn-config" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.812548 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6930d990-1321-4ec0-9422-91f1b5abbf5a" containerName="ovn-config" Feb 03 09:27:34 crc kubenswrapper[4756]: E0203 09:27:34.812565 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4bceb9-3d5a-4fe4-9b98-ed408d36116b" containerName="mariadb-account-create-update" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.812571 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4bceb9-3d5a-4fe4-9b98-ed408d36116b" containerName="mariadb-account-create-update" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.812717 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="6930d990-1321-4ec0-9422-91f1b5abbf5a" containerName="ovn-config" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.812736 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa4bceb9-3d5a-4fe4-9b98-ed408d36116b" containerName="mariadb-account-create-update" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.813225 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.815927 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.824581 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wnf95-config-p5bn8"] Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.926554 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-log-ovn\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.926616 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.926642 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq579\" (UniqueName: \"kubernetes.io/projected/2186a872-9466-4858-bfaa-46010072ff5b-kube-api-access-rq579\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.926691 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-scripts\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.926727 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-additional-scripts\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:34 crc kubenswrapper[4756]: I0203 09:27:34.926756 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run-ovn\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.028075 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-scripts\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.028218 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-additional-scripts\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.028264 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run-ovn\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.028306 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-log-ovn\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.028343 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.028371 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq579\" (UniqueName: \"kubernetes.io/projected/2186a872-9466-4858-bfaa-46010072ff5b-kube-api-access-rq579\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.030880 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-scripts\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.031291 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-additional-scripts\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.031584 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-log-ovn\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.031633 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run-ovn\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.031702 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.044420 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq579\" (UniqueName: \"kubernetes.io/projected/2186a872-9466-4858-bfaa-46010072ff5b-kube-api-access-rq579\") pod \"ovn-controller-wnf95-config-p5bn8\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.129861 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.569973 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wnf95-config-p5bn8"] Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.625527 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6930d990-1321-4ec0-9422-91f1b5abbf5a" path="/var/lib/kubelet/pods/6930d990-1321-4ec0-9422-91f1b5abbf5a/volumes" Feb 03 09:27:35 crc kubenswrapper[4756]: I0203 09:27:35.630794 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"f3885b181040fd27bffc7b541aa0cf0c0d75323e127933e5c35786474f80a7cb"} Feb 03 09:27:35 crc kubenswrapper[4756]: W0203 09:27:35.739102 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2186a872_9466_4858_bfaa_46010072ff5b.slice/crio-82c47a11b588877dc8dbae7b9688e81c4997fbbfb0ec653e9e70dd3e6a940836 WatchSource:0}: Error finding container 82c47a11b588877dc8dbae7b9688e81c4997fbbfb0ec653e9e70dd3e6a940836: Status 404 returned error can't find the container with id 82c47a11b588877dc8dbae7b9688e81c4997fbbfb0ec653e9e70dd3e6a940836 Feb 03 09:27:36 crc kubenswrapper[4756]: I0203 09:27:36.639664 4756 generic.go:334] "Generic (PLEG): container finished" podID="2186a872-9466-4858-bfaa-46010072ff5b" containerID="e2568a02083d69706f010c9cb5896146be34500a29706123dbde9047fa994f6f" exitCode=0 Feb 03 09:27:36 crc kubenswrapper[4756]: I0203 09:27:36.639759 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95-config-p5bn8" event={"ID":"2186a872-9466-4858-bfaa-46010072ff5b","Type":"ContainerDied","Data":"e2568a02083d69706f010c9cb5896146be34500a29706123dbde9047fa994f6f"} Feb 03 09:27:36 crc kubenswrapper[4756]: I0203 09:27:36.639967 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95-config-p5bn8" event={"ID":"2186a872-9466-4858-bfaa-46010072ff5b","Type":"ContainerStarted","Data":"82c47a11b588877dc8dbae7b9688e81c4997fbbfb0ec653e9e70dd3e6a940836"} Feb 03 09:27:36 crc kubenswrapper[4756]: I0203 09:27:36.652291 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"7147360d311f5d6b724e6bb76ceda21b9397832e034bc06607a974e6c16e7a78"} Feb 03 09:27:36 crc kubenswrapper[4756]: I0203 09:27:36.652519 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"93b087d4a69a2401ef6e6d3840042b25ee89e7d1c4cc833753273a05dcf99a82"} Feb 03 09:27:36 crc kubenswrapper[4756]: I0203 09:27:36.652650 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"4683d2353fd3b4d626eb349d3a340a1e24bbbf4093c666b40d063cd44166eeeb"} Feb 03 09:27:36 crc kubenswrapper[4756]: I0203 09:27:36.652746 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"ac87d1de7a731b5b613c80dcf295b85a4062c81a160b55b2dbf5106c632826c9"} Feb 03 09:27:37 crc kubenswrapper[4756]: I0203 09:27:37.665887 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"0ff50b3b61cb7b39e6ad29dc433596e3719fad04f995866ab8f981a133286a41"} Feb 03 09:27:37 crc kubenswrapper[4756]: I0203 09:27:37.666216 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"560f729f8fa288dc6baa1f25cf826167c9e5df094f5386f7e297edb187ff5183"} Feb 03 09:27:37 crc kubenswrapper[4756]: I0203 09:27:37.666230 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9","Type":"ContainerStarted","Data":"42aad02587409563fe1e93b24842725a5a92279e72078ca7f909b68c8dce4071"} Feb 03 09:27:37 crc kubenswrapper[4756]: I0203 09:27:37.699984 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.781098732 podStartE2EDuration="31.699957673s" podCreationTimestamp="2026-02-03 09:27:06 +0000 UTC" firstStartedPulling="2026-02-03 09:27:23.882383142 +0000 UTC m=+1035.032850517" lastFinishedPulling="2026-02-03 09:27:35.801242083 +0000 UTC m=+1046.951709458" observedRunningTime="2026-02-03 09:27:37.695663526 +0000 UTC m=+1048.846130921" watchObservedRunningTime="2026-02-03 09:27:37.699957673 +0000 UTC m=+1048.850425048" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.004887 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.066020 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-zqb5c"] Feb 03 09:27:38 crc kubenswrapper[4756]: E0203 09:27:38.066728 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2186a872-9466-4858-bfaa-46010072ff5b" containerName="ovn-config" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.066755 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="2186a872-9466-4858-bfaa-46010072ff5b" containerName="ovn-config" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.067024 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="2186a872-9466-4858-bfaa-46010072ff5b" containerName="ovn-config" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.068078 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.070156 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.074196 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-zqb5c"] Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091290 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run\") pod \"2186a872-9466-4858-bfaa-46010072ff5b\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091348 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-additional-scripts\") pod \"2186a872-9466-4858-bfaa-46010072ff5b\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091383 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run-ovn\") pod \"2186a872-9466-4858-bfaa-46010072ff5b\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091404 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-log-ovn\") pod \"2186a872-9466-4858-bfaa-46010072ff5b\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091525 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq579\" (UniqueName: \"kubernetes.io/projected/2186a872-9466-4858-bfaa-46010072ff5b-kube-api-access-rq579\") pod \"2186a872-9466-4858-bfaa-46010072ff5b\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091546 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-scripts\") pod \"2186a872-9466-4858-bfaa-46010072ff5b\" (UID: \"2186a872-9466-4858-bfaa-46010072ff5b\") " Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091769 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-config\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091809 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091827 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091861 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091902 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.091921 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qm4q\" (UniqueName: \"kubernetes.io/projected/35741ef2-602f-4fa0-9a37-d430ddc73dfe-kube-api-access-4qm4q\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.092009 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run" (OuterVolumeSpecName: "var-run") pod "2186a872-9466-4858-bfaa-46010072ff5b" (UID: "2186a872-9466-4858-bfaa-46010072ff5b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.092705 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "2186a872-9466-4858-bfaa-46010072ff5b" (UID: "2186a872-9466-4858-bfaa-46010072ff5b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.092735 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "2186a872-9466-4858-bfaa-46010072ff5b" (UID: "2186a872-9466-4858-bfaa-46010072ff5b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.093487 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-scripts" (OuterVolumeSpecName: "scripts") pod "2186a872-9466-4858-bfaa-46010072ff5b" (UID: "2186a872-9466-4858-bfaa-46010072ff5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.093523 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "2186a872-9466-4858-bfaa-46010072ff5b" (UID: "2186a872-9466-4858-bfaa-46010072ff5b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.097827 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2186a872-9466-4858-bfaa-46010072ff5b-kube-api-access-rq579" (OuterVolumeSpecName: "kube-api-access-rq579") pod "2186a872-9466-4858-bfaa-46010072ff5b" (UID: "2186a872-9466-4858-bfaa-46010072ff5b"). InnerVolumeSpecName "kube-api-access-rq579". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193335 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193384 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193431 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193492 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qm4q\" (UniqueName: \"kubernetes.io/projected/35741ef2-602f-4fa0-9a37-d430ddc73dfe-kube-api-access-4qm4q\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193513 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193571 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-config\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193628 4756 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193640 4756 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193648 4756 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193657 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq579\" (UniqueName: \"kubernetes.io/projected/2186a872-9466-4858-bfaa-46010072ff5b-kube-api-access-rq579\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193667 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2186a872-9466-4858-bfaa-46010072ff5b-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.193676 4756 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2186a872-9466-4858-bfaa-46010072ff5b-var-run\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.194643 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.194675 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.194684 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-config\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.194793 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.195357 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.215683 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qm4q\" (UniqueName: \"kubernetes.io/projected/35741ef2-602f-4fa0-9a37-d430ddc73dfe-kube-api-access-4qm4q\") pod \"dnsmasq-dns-5c79d794d7-zqb5c\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.388497 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.674980 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wnf95-config-p5bn8" event={"ID":"2186a872-9466-4858-bfaa-46010072ff5b","Type":"ContainerDied","Data":"82c47a11b588877dc8dbae7b9688e81c4997fbbfb0ec653e9e70dd3e6a940836"} Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.675074 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82c47a11b588877dc8dbae7b9688e81c4997fbbfb0ec653e9e70dd3e6a940836" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.675106 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wnf95-config-p5bn8" Feb 03 09:27:38 crc kubenswrapper[4756]: I0203 09:27:38.902258 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-zqb5c"] Feb 03 09:27:38 crc kubenswrapper[4756]: W0203 09:27:38.902773 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35741ef2_602f_4fa0_9a37_d430ddc73dfe.slice/crio-b7f88e1033c5f71f2e6b68f0bf49c3c4ad406c2516927a7f9bb79c6bc002c24e WatchSource:0}: Error finding container b7f88e1033c5f71f2e6b68f0bf49c3c4ad406c2516927a7f9bb79c6bc002c24e: Status 404 returned error can't find the container with id b7f88e1033c5f71f2e6b68f0bf49c3c4ad406c2516927a7f9bb79c6bc002c24e Feb 03 09:27:39 crc kubenswrapper[4756]: I0203 09:27:39.086129 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-wnf95-config-p5bn8"] Feb 03 09:27:39 crc kubenswrapper[4756]: I0203 09:27:39.093808 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-wnf95-config-p5bn8"] Feb 03 09:27:39 crc kubenswrapper[4756]: I0203 09:27:39.622856 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2186a872-9466-4858-bfaa-46010072ff5b" path="/var/lib/kubelet/pods/2186a872-9466-4858-bfaa-46010072ff5b/volumes" Feb 03 09:27:39 crc kubenswrapper[4756]: I0203 09:27:39.689378 4756 generic.go:334] "Generic (PLEG): container finished" podID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" containerID="3f0c9e8eed23f0039ec52218af552fd2a58f784a32516cdfd709249f61f9c893" exitCode=0 Feb 03 09:27:39 crc kubenswrapper[4756]: I0203 09:27:39.689426 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" event={"ID":"35741ef2-602f-4fa0-9a37-d430ddc73dfe","Type":"ContainerDied","Data":"3f0c9e8eed23f0039ec52218af552fd2a58f784a32516cdfd709249f61f9c893"} Feb 03 09:27:39 crc kubenswrapper[4756]: I0203 09:27:39.689460 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" event={"ID":"35741ef2-602f-4fa0-9a37-d430ddc73dfe","Type":"ContainerStarted","Data":"b7f88e1033c5f71f2e6b68f0bf49c3c4ad406c2516927a7f9bb79c6bc002c24e"} Feb 03 09:27:40 crc kubenswrapper[4756]: I0203 09:27:40.690678 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 03 09:27:40 crc kubenswrapper[4756]: I0203 09:27:40.699246 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" event={"ID":"35741ef2-602f-4fa0-9a37-d430ddc73dfe","Type":"ContainerStarted","Data":"49aeb70a9162f4f3229b96217718fbddaaf1cccd200f03eaf94fbb74be2ac6d0"} Feb 03 09:27:40 crc kubenswrapper[4756]: I0203 09:27:40.699573 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.021671 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.026914 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" podStartSLOduration=3.026892205 podStartE2EDuration="3.026892205s" podCreationTimestamp="2026-02-03 09:27:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:40.760934385 +0000 UTC m=+1051.911401770" watchObservedRunningTime="2026-02-03 09:27:41.026892205 +0000 UTC m=+1052.177359580" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.029667 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-jbrwj"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.030952 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.038634 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jbrwj"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.160677 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e555442-e24a-467c-8ab3-426856e59e43-operator-scripts\") pod \"cinder-db-create-jbrwj\" (UID: \"7e555442-e24a-467c-8ab3-426856e59e43\") " pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.160743 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hh62\" (UniqueName: \"kubernetes.io/projected/7e555442-e24a-467c-8ab3-426856e59e43-kube-api-access-4hh62\") pod \"cinder-db-create-jbrwj\" (UID: \"7e555442-e24a-467c-8ab3-426856e59e43\") " pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.167394 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-b948-account-create-update-q4k64"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.168688 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.174746 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.179052 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-nz7sg"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.180692 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.188266 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nz7sg"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.202326 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b948-account-create-update-q4k64"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.251917 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-88aa-account-create-update-ndhz5"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.253102 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.254842 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.262516 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkr5r\" (UniqueName: \"kubernetes.io/projected/0f0bf352-90f0-4613-b6c8-9206ddbc4235-kube-api-access-bkr5r\") pod \"cinder-b948-account-create-update-q4k64\" (UID: \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\") " pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.262622 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f0bf352-90f0-4613-b6c8-9206ddbc4235-operator-scripts\") pod \"cinder-b948-account-create-update-q4k64\" (UID: \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\") " pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.262669 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e555442-e24a-467c-8ab3-426856e59e43-operator-scripts\") pod \"cinder-db-create-jbrwj\" (UID: \"7e555442-e24a-467c-8ab3-426856e59e43\") " pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.262697 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hh62\" (UniqueName: \"kubernetes.io/projected/7e555442-e24a-467c-8ab3-426856e59e43-kube-api-access-4hh62\") pod \"cinder-db-create-jbrwj\" (UID: \"7e555442-e24a-467c-8ab3-426856e59e43\") " pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.263711 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e555442-e24a-467c-8ab3-426856e59e43-operator-scripts\") pod \"cinder-db-create-jbrwj\" (UID: \"7e555442-e24a-467c-8ab3-426856e59e43\") " pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.265036 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-88aa-account-create-update-ndhz5"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.298351 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hh62\" (UniqueName: \"kubernetes.io/projected/7e555442-e24a-467c-8ab3-426856e59e43-kube-api-access-4hh62\") pod \"cinder-db-create-jbrwj\" (UID: \"7e555442-e24a-467c-8ab3-426856e59e43\") " pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.328247 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-h5vn4"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.329271 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.331526 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fh27s" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.331736 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.331880 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.332044 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.344974 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-h5vn4"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.349142 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.367998 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdc2s\" (UniqueName: \"kubernetes.io/projected/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-kube-api-access-bdc2s\") pod \"barbican-88aa-account-create-update-ndhz5\" (UID: \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\") " pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.368077 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnprm\" (UniqueName: \"kubernetes.io/projected/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-kube-api-access-tnprm\") pod \"barbican-db-create-nz7sg\" (UID: \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\") " pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.368127 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkr5r\" (UniqueName: \"kubernetes.io/projected/0f0bf352-90f0-4613-b6c8-9206ddbc4235-kube-api-access-bkr5r\") pod \"cinder-b948-account-create-update-q4k64\" (UID: \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\") " pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.368168 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-operator-scripts\") pod \"barbican-db-create-nz7sg\" (UID: \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\") " pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.368216 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-operator-scripts\") pod \"barbican-88aa-account-create-update-ndhz5\" (UID: \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\") " pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.368247 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f0bf352-90f0-4613-b6c8-9206ddbc4235-operator-scripts\") pod \"cinder-b948-account-create-update-q4k64\" (UID: \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\") " pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.369192 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f0bf352-90f0-4613-b6c8-9206ddbc4235-operator-scripts\") pod \"cinder-b948-account-create-update-q4k64\" (UID: \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\") " pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.395413 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkr5r\" (UniqueName: \"kubernetes.io/projected/0f0bf352-90f0-4613-b6c8-9206ddbc4235-kube-api-access-bkr5r\") pod \"cinder-b948-account-create-update-q4k64\" (UID: \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\") " pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.454762 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-hgsfq"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.456039 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.463912 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hgsfq"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.469391 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lvg5\" (UniqueName: \"kubernetes.io/projected/3e62dcdb-516f-406c-a6dd-d780f6c250bc-kube-api-access-5lvg5\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.469457 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-config-data\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.469509 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdc2s\" (UniqueName: \"kubernetes.io/projected/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-kube-api-access-bdc2s\") pod \"barbican-88aa-account-create-update-ndhz5\" (UID: \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\") " pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.469553 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnprm\" (UniqueName: \"kubernetes.io/projected/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-kube-api-access-tnprm\") pod \"barbican-db-create-nz7sg\" (UID: \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\") " pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.469617 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-combined-ca-bundle\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.469643 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-operator-scripts\") pod \"barbican-db-create-nz7sg\" (UID: \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\") " pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.469731 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-operator-scripts\") pod \"barbican-88aa-account-create-update-ndhz5\" (UID: \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\") " pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.471124 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-operator-scripts\") pod \"barbican-db-create-nz7sg\" (UID: \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\") " pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.471653 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-operator-scripts\") pod \"barbican-88aa-account-create-update-ndhz5\" (UID: \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\") " pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.491086 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdc2s\" (UniqueName: \"kubernetes.io/projected/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-kube-api-access-bdc2s\") pod \"barbican-88aa-account-create-update-ndhz5\" (UID: \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\") " pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.494079 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.498963 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnprm\" (UniqueName: \"kubernetes.io/projected/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-kube-api-access-tnprm\") pod \"barbican-db-create-nz7sg\" (UID: \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\") " pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.514981 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.564419 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-953f-account-create-update-595f4"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.565682 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.569941 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.571230 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk4rh\" (UniqueName: \"kubernetes.io/projected/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-kube-api-access-vk4rh\") pod \"neutron-db-create-hgsfq\" (UID: \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\") " pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.571323 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lvg5\" (UniqueName: \"kubernetes.io/projected/3e62dcdb-516f-406c-a6dd-d780f6c250bc-kube-api-access-5lvg5\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.571348 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-operator-scripts\") pod \"neutron-db-create-hgsfq\" (UID: \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\") " pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.571368 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-config-data\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.571473 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-combined-ca-bundle\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.575181 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-config-data\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.578183 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-combined-ca-bundle\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.581868 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.589957 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-953f-account-create-update-595f4"] Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.595770 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lvg5\" (UniqueName: \"kubernetes.io/projected/3e62dcdb-516f-406c-a6dd-d780f6c250bc-kube-api-access-5lvg5\") pod \"keystone-db-sync-h5vn4\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.652663 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.674234 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk4rh\" (UniqueName: \"kubernetes.io/projected/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-kube-api-access-vk4rh\") pod \"neutron-db-create-hgsfq\" (UID: \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\") " pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.674345 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-operator-scripts\") pod \"neutron-db-create-hgsfq\" (UID: \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\") " pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.674425 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr9z8\" (UniqueName: \"kubernetes.io/projected/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-kube-api-access-hr9z8\") pod \"neutron-953f-account-create-update-595f4\" (UID: \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\") " pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.674511 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-operator-scripts\") pod \"neutron-953f-account-create-update-595f4\" (UID: \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\") " pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.675382 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-operator-scripts\") pod \"neutron-db-create-hgsfq\" (UID: \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\") " pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.719923 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk4rh\" (UniqueName: \"kubernetes.io/projected/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-kube-api-access-vk4rh\") pod \"neutron-db-create-hgsfq\" (UID: \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\") " pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.734892 4756 generic.go:334] "Generic (PLEG): container finished" podID="d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" containerID="57832a58c9733cca830b6ae89f7f260772e41d1643546ecd029a943252c27533" exitCode=0 Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.735005 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8j6l5" event={"ID":"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9","Type":"ContainerDied","Data":"57832a58c9733cca830b6ae89f7f260772e41d1643546ecd029a943252c27533"} Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.776636 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr9z8\" (UniqueName: \"kubernetes.io/projected/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-kube-api-access-hr9z8\") pod \"neutron-953f-account-create-update-595f4\" (UID: \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\") " pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.776706 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-operator-scripts\") pod \"neutron-953f-account-create-update-595f4\" (UID: \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\") " pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.779704 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-operator-scripts\") pod \"neutron-953f-account-create-update-595f4\" (UID: \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\") " pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.781244 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.806921 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr9z8\" (UniqueName: \"kubernetes.io/projected/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-kube-api-access-hr9z8\") pod \"neutron-953f-account-create-update-595f4\" (UID: \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\") " pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.972056 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:41 crc kubenswrapper[4756]: I0203 09:27:41.994986 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jbrwj"] Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.192825 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b948-account-create-update-q4k64"] Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.244484 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-h5vn4"] Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.274780 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-88aa-account-create-update-ndhz5"] Feb 03 09:27:42 crc kubenswrapper[4756]: W0203 09:27:42.297704 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad7fc552_1c39_4c32_a9ea_7df1556ce73f.slice/crio-444a13afdc03060fccba0589ed5878a47735e2e018edc0ec8397774edbb13c39 WatchSource:0}: Error finding container 444a13afdc03060fccba0589ed5878a47735e2e018edc0ec8397774edbb13c39: Status 404 returned error can't find the container with id 444a13afdc03060fccba0589ed5878a47735e2e018edc0ec8397774edbb13c39 Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.299495 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nz7sg"] Feb 03 09:27:42 crc kubenswrapper[4756]: W0203 09:27:42.330506 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4e6768f_7cd7_4a37_ab83_59b8fa562c11.slice/crio-6cef8d19c2087835afce157b669b8c6bf91b45db02614c74498f5ee69e0ac628 WatchSource:0}: Error finding container 6cef8d19c2087835afce157b669b8c6bf91b45db02614c74498f5ee69e0ac628: Status 404 returned error can't find the container with id 6cef8d19c2087835afce157b669b8c6bf91b45db02614c74498f5ee69e0ac628 Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.498354 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hgsfq"] Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.596391 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-953f-account-create-update-595f4"] Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.743593 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-h5vn4" event={"ID":"3e62dcdb-516f-406c-a6dd-d780f6c250bc","Type":"ContainerStarted","Data":"bb447635f1a9aa81b44e53aec1860c5a1f5669a8dc2bbb4d1660637c2744d25e"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.747102 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nz7sg" event={"ID":"a4e6768f-7cd7-4a37-ab83-59b8fa562c11","Type":"ContainerStarted","Data":"6cef8d19c2087835afce157b669b8c6bf91b45db02614c74498f5ee69e0ac628"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.748500 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-953f-account-create-update-595f4" event={"ID":"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051","Type":"ContainerStarted","Data":"0512771fbd6232f3e7263c742b8d671f8c340a401fd100eff0d41be7af660582"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.750504 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b948-account-create-update-q4k64" event={"ID":"0f0bf352-90f0-4613-b6c8-9206ddbc4235","Type":"ContainerStarted","Data":"896fafbfcde87e7c0add95627b1ce3d650a8c3ebfaf4f765df460f2347945fbd"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.750561 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b948-account-create-update-q4k64" event={"ID":"0f0bf352-90f0-4613-b6c8-9206ddbc4235","Type":"ContainerStarted","Data":"8051930d46bbd871ebd5641a53a72f1fde6dc3d95fc45deef65dc9261ab97164"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.752309 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hgsfq" event={"ID":"5971a24c-30d3-4e3e-a649-6d6f637dbfd5","Type":"ContainerStarted","Data":"461d47bb89ec1a340dba420aa36489f917e54b06014617f300ea3aca3fa8dbaa"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.754364 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jbrwj" event={"ID":"7e555442-e24a-467c-8ab3-426856e59e43","Type":"ContainerStarted","Data":"8ca8f574bcef2d1e934268f84949aeeeba5ee177f6c8ef940f7cbd5ca210ee30"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.754421 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jbrwj" event={"ID":"7e555442-e24a-467c-8ab3-426856e59e43","Type":"ContainerStarted","Data":"44ad644e70e13dc4163443a79033a6da63890cae5e3f2a870851b8883d4d68dd"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.755829 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-88aa-account-create-update-ndhz5" event={"ID":"ad7fc552-1c39-4c32-a9ea-7df1556ce73f","Type":"ContainerStarted","Data":"444a13afdc03060fccba0589ed5878a47735e2e018edc0ec8397774edbb13c39"} Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.791790 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-b948-account-create-update-q4k64" podStartSLOduration=1.7917525269999999 podStartE2EDuration="1.791752527s" podCreationTimestamp="2026-02-03 09:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:42.767590719 +0000 UTC m=+1053.918058104" watchObservedRunningTime="2026-02-03 09:27:42.791752527 +0000 UTC m=+1053.942219912" Feb 03 09:27:42 crc kubenswrapper[4756]: I0203 09:27:42.795242 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-jbrwj" podStartSLOduration=1.7952332279999998 podStartE2EDuration="1.795233228s" podCreationTimestamp="2026-02-03 09:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:42.791161768 +0000 UTC m=+1053.941629143" watchObservedRunningTime="2026-02-03 09:27:42.795233228 +0000 UTC m=+1053.945700603" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.160089 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.306005 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-db-sync-config-data\") pod \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.306374 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq7nr\" (UniqueName: \"kubernetes.io/projected/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-kube-api-access-kq7nr\") pod \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.306413 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-combined-ca-bundle\") pod \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.306478 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-config-data\") pod \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\" (UID: \"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9\") " Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.314783 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" (UID: "d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.315197 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-kube-api-access-kq7nr" (OuterVolumeSpecName: "kube-api-access-kq7nr") pod "d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" (UID: "d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9"). InnerVolumeSpecName "kube-api-access-kq7nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.333869 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" (UID: "d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.364434 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-config-data" (OuterVolumeSpecName: "config-data") pod "d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" (UID: "d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.408792 4756 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.408843 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq7nr\" (UniqueName: \"kubernetes.io/projected/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-kube-api-access-kq7nr\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.408857 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.408871 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.794144 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8j6l5" Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.794657 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8j6l5" event={"ID":"d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9","Type":"ContainerDied","Data":"33938a62e18a373c2dae7174343ee1d4757cb94fb90d209e0acd3f66f5ccb7af"} Feb 03 09:27:43 crc kubenswrapper[4756]: I0203 09:27:43.794682 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33938a62e18a373c2dae7174343ee1d4757cb94fb90d209e0acd3f66f5ccb7af" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.077181 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-zqb5c"] Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.077686 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" podUID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" containerName="dnsmasq-dns" containerID="cri-o://49aeb70a9162f4f3229b96217718fbddaaf1cccd200f03eaf94fbb74be2ac6d0" gracePeriod=10 Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.095919 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.208391 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-vmwts"] Feb 03 09:27:44 crc kubenswrapper[4756]: E0203 09:27:44.208870 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" containerName="glance-db-sync" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.208895 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" containerName="glance-db-sync" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.209086 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" containerName="glance-db-sync" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.222085 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.226865 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-vmwts"] Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.327937 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldtbs\" (UniqueName: \"kubernetes.io/projected/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-kube-api-access-ldtbs\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.327980 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.328017 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.328051 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-config\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.328084 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.328111 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.429122 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.429194 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-config\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.429226 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.429258 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.429329 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldtbs\" (UniqueName: \"kubernetes.io/projected/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-kube-api-access-ldtbs\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.429348 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.430294 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.430325 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.430370 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-config\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.430336 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.431088 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.451870 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldtbs\" (UniqueName: \"kubernetes.io/projected/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-kube-api-access-ldtbs\") pod \"dnsmasq-dns-5f59b8f679-vmwts\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:44 crc kubenswrapper[4756]: I0203 09:27:44.582568 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.064296 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-vmwts"] Feb 03 09:27:45 crc kubenswrapper[4756]: W0203 09:27:45.069877 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c1ee136_d3b1_4d03_8313_81dd9c0f882b.slice/crio-b500f8c760b3bc134764b66bea2156424566f48160f8bec8726af5ddf01dd6ed WatchSource:0}: Error finding container b500f8c760b3bc134764b66bea2156424566f48160f8bec8726af5ddf01dd6ed: Status 404 returned error can't find the container with id b500f8c760b3bc134764b66bea2156424566f48160f8bec8726af5ddf01dd6ed Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.816156 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hgsfq" event={"ID":"5971a24c-30d3-4e3e-a649-6d6f637dbfd5","Type":"ContainerStarted","Data":"ab54b30466202333f111da6ab5eaae031f752ead2870986ae79ef42231ddb2ed"} Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.819893 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-88aa-account-create-update-ndhz5" event={"ID":"ad7fc552-1c39-4c32-a9ea-7df1556ce73f","Type":"ContainerStarted","Data":"14aca9176edc225e01b661d9c251b5f3be6bd3785c25e8068a9c1eec85d47e16"} Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.822165 4756 generic.go:334] "Generic (PLEG): container finished" podID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" containerID="49aeb70a9162f4f3229b96217718fbddaaf1cccd200f03eaf94fbb74be2ac6d0" exitCode=0 Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.822216 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" event={"ID":"35741ef2-602f-4fa0-9a37-d430ddc73dfe","Type":"ContainerDied","Data":"49aeb70a9162f4f3229b96217718fbddaaf1cccd200f03eaf94fbb74be2ac6d0"} Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.823930 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nz7sg" event={"ID":"a4e6768f-7cd7-4a37-ab83-59b8fa562c11","Type":"ContainerStarted","Data":"ae612df2dc0b954d698378e015dc7145ec044a37df2bf55abf1e2455834b5982"} Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.826655 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-953f-account-create-update-595f4" event={"ID":"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051","Type":"ContainerStarted","Data":"3dddb67456e4c4f60bdd8c77ab68cab576f8d923ec669b7642248cc8b0cc0057"} Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.828893 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" event={"ID":"3c1ee136-d3b1-4d03-8313-81dd9c0f882b","Type":"ContainerStarted","Data":"2aa552b13be2cc0ddb763534153a4d76f01f5fe14d976094d8381330f8593b2d"} Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.828930 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" event={"ID":"3c1ee136-d3b1-4d03-8313-81dd9c0f882b","Type":"ContainerStarted","Data":"b500f8c760b3bc134764b66bea2156424566f48160f8bec8726af5ddf01dd6ed"} Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.835457 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-hgsfq" podStartSLOduration=4.835423069 podStartE2EDuration="4.835423069s" podCreationTimestamp="2026-02-03 09:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:45.831974419 +0000 UTC m=+1056.982441794" watchObservedRunningTime="2026-02-03 09:27:45.835423069 +0000 UTC m=+1056.985890444" Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.866798 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-88aa-account-create-update-ndhz5" podStartSLOduration=4.866780607 podStartE2EDuration="4.866780607s" podCreationTimestamp="2026-02-03 09:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:45.859103342 +0000 UTC m=+1057.009570717" watchObservedRunningTime="2026-02-03 09:27:45.866780607 +0000 UTC m=+1057.017247982" Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.916394 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-nz7sg" podStartSLOduration=4.916371994 podStartE2EDuration="4.916371994s" podCreationTimestamp="2026-02-03 09:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:45.906461168 +0000 UTC m=+1057.056928553" watchObservedRunningTime="2026-02-03 09:27:45.916371994 +0000 UTC m=+1057.066839369" Feb 03 09:27:45 crc kubenswrapper[4756]: I0203 09:27:45.932335 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-953f-account-create-update-595f4" podStartSLOduration=4.932309071 podStartE2EDuration="4.932309071s" podCreationTimestamp="2026-02-03 09:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:45.925637459 +0000 UTC m=+1057.076104834" watchObservedRunningTime="2026-02-03 09:27:45.932309071 +0000 UTC m=+1057.082776446" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.327684 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.457555 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-sb\") pod \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.457653 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qm4q\" (UniqueName: \"kubernetes.io/projected/35741ef2-602f-4fa0-9a37-d430ddc73dfe-kube-api-access-4qm4q\") pod \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.457694 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-swift-storage-0\") pod \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.457805 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-config\") pod \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.457832 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-nb\") pod \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.457857 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-svc\") pod \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\" (UID: \"35741ef2-602f-4fa0-9a37-d430ddc73dfe\") " Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.464355 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35741ef2-602f-4fa0-9a37-d430ddc73dfe-kube-api-access-4qm4q" (OuterVolumeSpecName: "kube-api-access-4qm4q") pod "35741ef2-602f-4fa0-9a37-d430ddc73dfe" (UID: "35741ef2-602f-4fa0-9a37-d430ddc73dfe"). InnerVolumeSpecName "kube-api-access-4qm4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.504112 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35741ef2-602f-4fa0-9a37-d430ddc73dfe" (UID: "35741ef2-602f-4fa0-9a37-d430ddc73dfe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.504900 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-config" (OuterVolumeSpecName: "config") pod "35741ef2-602f-4fa0-9a37-d430ddc73dfe" (UID: "35741ef2-602f-4fa0-9a37-d430ddc73dfe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.508141 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "35741ef2-602f-4fa0-9a37-d430ddc73dfe" (UID: "35741ef2-602f-4fa0-9a37-d430ddc73dfe"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.518313 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35741ef2-602f-4fa0-9a37-d430ddc73dfe" (UID: "35741ef2-602f-4fa0-9a37-d430ddc73dfe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.527663 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35741ef2-602f-4fa0-9a37-d430ddc73dfe" (UID: "35741ef2-602f-4fa0-9a37-d430ddc73dfe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.560894 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.560937 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qm4q\" (UniqueName: \"kubernetes.io/projected/35741ef2-602f-4fa0-9a37-d430ddc73dfe-kube-api-access-4qm4q\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.560954 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.560969 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.561015 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.561027 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35741ef2-602f-4fa0-9a37-d430ddc73dfe-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.839245 4756 generic.go:334] "Generic (PLEG): container finished" podID="0f0bf352-90f0-4613-b6c8-9206ddbc4235" containerID="896fafbfcde87e7c0add95627b1ce3d650a8c3ebfaf4f765df460f2347945fbd" exitCode=0 Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.839349 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b948-account-create-update-q4k64" event={"ID":"0f0bf352-90f0-4613-b6c8-9206ddbc4235","Type":"ContainerDied","Data":"896fafbfcde87e7c0add95627b1ce3d650a8c3ebfaf4f765df460f2347945fbd"} Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.843388 4756 generic.go:334] "Generic (PLEG): container finished" podID="5971a24c-30d3-4e3e-a649-6d6f637dbfd5" containerID="ab54b30466202333f111da6ab5eaae031f752ead2870986ae79ef42231ddb2ed" exitCode=0 Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.843461 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hgsfq" event={"ID":"5971a24c-30d3-4e3e-a649-6d6f637dbfd5","Type":"ContainerDied","Data":"ab54b30466202333f111da6ab5eaae031f752ead2870986ae79ef42231ddb2ed"} Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.845515 4756 generic.go:334] "Generic (PLEG): container finished" podID="7e555442-e24a-467c-8ab3-426856e59e43" containerID="8ca8f574bcef2d1e934268f84949aeeeba5ee177f6c8ef940f7cbd5ca210ee30" exitCode=0 Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.845556 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jbrwj" event={"ID":"7e555442-e24a-467c-8ab3-426856e59e43","Type":"ContainerDied","Data":"8ca8f574bcef2d1e934268f84949aeeeba5ee177f6c8ef940f7cbd5ca210ee30"} Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.847157 4756 generic.go:334] "Generic (PLEG): container finished" podID="ad7fc552-1c39-4c32-a9ea-7df1556ce73f" containerID="14aca9176edc225e01b661d9c251b5f3be6bd3785c25e8068a9c1eec85d47e16" exitCode=0 Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.847210 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-88aa-account-create-update-ndhz5" event={"ID":"ad7fc552-1c39-4c32-a9ea-7df1556ce73f","Type":"ContainerDied","Data":"14aca9176edc225e01b661d9c251b5f3be6bd3785c25e8068a9c1eec85d47e16"} Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.850058 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.850067 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-zqb5c" event={"ID":"35741ef2-602f-4fa0-9a37-d430ddc73dfe","Type":"ContainerDied","Data":"b7f88e1033c5f71f2e6b68f0bf49c3c4ad406c2516927a7f9bb79c6bc002c24e"} Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.850132 4756 scope.go:117] "RemoveContainer" containerID="49aeb70a9162f4f3229b96217718fbddaaf1cccd200f03eaf94fbb74be2ac6d0" Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.853059 4756 generic.go:334] "Generic (PLEG): container finished" podID="a4e6768f-7cd7-4a37-ab83-59b8fa562c11" containerID="ae612df2dc0b954d698378e015dc7145ec044a37df2bf55abf1e2455834b5982" exitCode=0 Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.853158 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nz7sg" event={"ID":"a4e6768f-7cd7-4a37-ab83-59b8fa562c11","Type":"ContainerDied","Data":"ae612df2dc0b954d698378e015dc7145ec044a37df2bf55abf1e2455834b5982"} Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.855618 4756 generic.go:334] "Generic (PLEG): container finished" podID="e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051" containerID="3dddb67456e4c4f60bdd8c77ab68cab576f8d923ec669b7642248cc8b0cc0057" exitCode=0 Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.855876 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-953f-account-create-update-595f4" event={"ID":"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051","Type":"ContainerDied","Data":"3dddb67456e4c4f60bdd8c77ab68cab576f8d923ec669b7642248cc8b0cc0057"} Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.857555 4756 generic.go:334] "Generic (PLEG): container finished" podID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerID="2aa552b13be2cc0ddb763534153a4d76f01f5fe14d976094d8381330f8593b2d" exitCode=0 Feb 03 09:27:46 crc kubenswrapper[4756]: I0203 09:27:46.857614 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" event={"ID":"3c1ee136-d3b1-4d03-8313-81dd9c0f882b","Type":"ContainerDied","Data":"2aa552b13be2cc0ddb763534153a4d76f01f5fe14d976094d8381330f8593b2d"} Feb 03 09:27:47 crc kubenswrapper[4756]: I0203 09:27:47.002563 4756 scope.go:117] "RemoveContainer" containerID="3f0c9e8eed23f0039ec52218af552fd2a58f784a32516cdfd709249f61f9c893" Feb 03 09:27:47 crc kubenswrapper[4756]: I0203 09:27:47.026591 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-zqb5c"] Feb 03 09:27:47 crc kubenswrapper[4756]: I0203 09:27:47.033523 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-zqb5c"] Feb 03 09:27:47 crc kubenswrapper[4756]: I0203 09:27:47.625564 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" path="/var/lib/kubelet/pods/35741ef2-602f-4fa0-9a37-d430ddc73dfe/volumes" Feb 03 09:27:47 crc kubenswrapper[4756]: I0203 09:27:47.868700 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" event={"ID":"3c1ee136-d3b1-4d03-8313-81dd9c0f882b","Type":"ContainerStarted","Data":"10211dd17ec07922acf6c530bd708fb3d69f765c93b273fe332609b84fcb8994"} Feb 03 09:27:47 crc kubenswrapper[4756]: I0203 09:27:47.868992 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:47 crc kubenswrapper[4756]: I0203 09:27:47.899605 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" podStartSLOduration=3.8995867029999998 podStartE2EDuration="3.899586703s" podCreationTimestamp="2026-02-03 09:27:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:27:47.899246042 +0000 UTC m=+1059.049713417" watchObservedRunningTime="2026-02-03 09:27:47.899586703 +0000 UTC m=+1059.050054078" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.599334 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.605661 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.623015 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.644517 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.659022 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.677157 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.740997 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-operator-scripts\") pod \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\" (UID: \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741065 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-operator-scripts\") pod \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\" (UID: \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741100 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e555442-e24a-467c-8ab3-426856e59e43-operator-scripts\") pod \"7e555442-e24a-467c-8ab3-426856e59e43\" (UID: \"7e555442-e24a-467c-8ab3-426856e59e43\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741122 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnprm\" (UniqueName: \"kubernetes.io/projected/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-kube-api-access-tnprm\") pod \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\" (UID: \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741150 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdc2s\" (UniqueName: \"kubernetes.io/projected/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-kube-api-access-bdc2s\") pod \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\" (UID: \"ad7fc552-1c39-4c32-a9ea-7df1556ce73f\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741192 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-operator-scripts\") pod \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\" (UID: \"a4e6768f-7cd7-4a37-ab83-59b8fa562c11\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741220 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr9z8\" (UniqueName: \"kubernetes.io/projected/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-kube-api-access-hr9z8\") pod \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\" (UID: \"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741296 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkr5r\" (UniqueName: \"kubernetes.io/projected/0f0bf352-90f0-4613-b6c8-9206ddbc4235-kube-api-access-bkr5r\") pod \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\" (UID: \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741313 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f0bf352-90f0-4613-b6c8-9206ddbc4235-operator-scripts\") pod \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\" (UID: \"0f0bf352-90f0-4613-b6c8-9206ddbc4235\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.741378 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hh62\" (UniqueName: \"kubernetes.io/projected/7e555442-e24a-467c-8ab3-426856e59e43-kube-api-access-4hh62\") pod \"7e555442-e24a-467c-8ab3-426856e59e43\" (UID: \"7e555442-e24a-467c-8ab3-426856e59e43\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.742947 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051" (UID: "e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.742990 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f0bf352-90f0-4613-b6c8-9206ddbc4235-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f0bf352-90f0-4613-b6c8-9206ddbc4235" (UID: "0f0bf352-90f0-4613-b6c8-9206ddbc4235"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.743015 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a4e6768f-7cd7-4a37-ab83-59b8fa562c11" (UID: "a4e6768f-7cd7-4a37-ab83-59b8fa562c11"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.743410 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad7fc552-1c39-4c32-a9ea-7df1556ce73f" (UID: "ad7fc552-1c39-4c32-a9ea-7df1556ce73f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.743428 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e555442-e24a-467c-8ab3-426856e59e43-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e555442-e24a-467c-8ab3-426856e59e43" (UID: "7e555442-e24a-467c-8ab3-426856e59e43"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.747316 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e555442-e24a-467c-8ab3-426856e59e43-kube-api-access-4hh62" (OuterVolumeSpecName: "kube-api-access-4hh62") pod "7e555442-e24a-467c-8ab3-426856e59e43" (UID: "7e555442-e24a-467c-8ab3-426856e59e43"). InnerVolumeSpecName "kube-api-access-4hh62". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.748001 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-kube-api-access-hr9z8" (OuterVolumeSpecName: "kube-api-access-hr9z8") pod "e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051" (UID: "e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051"). InnerVolumeSpecName "kube-api-access-hr9z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.748113 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-kube-api-access-tnprm" (OuterVolumeSpecName: "kube-api-access-tnprm") pod "a4e6768f-7cd7-4a37-ab83-59b8fa562c11" (UID: "a4e6768f-7cd7-4a37-ab83-59b8fa562c11"). InnerVolumeSpecName "kube-api-access-tnprm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.748593 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0bf352-90f0-4613-b6c8-9206ddbc4235-kube-api-access-bkr5r" (OuterVolumeSpecName: "kube-api-access-bkr5r") pod "0f0bf352-90f0-4613-b6c8-9206ddbc4235" (UID: "0f0bf352-90f0-4613-b6c8-9206ddbc4235"). InnerVolumeSpecName "kube-api-access-bkr5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.757019 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-kube-api-access-bdc2s" (OuterVolumeSpecName: "kube-api-access-bdc2s") pod "ad7fc552-1c39-4c32-a9ea-7df1556ce73f" (UID: "ad7fc552-1c39-4c32-a9ea-7df1556ce73f"). InnerVolumeSpecName "kube-api-access-bdc2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.842756 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-operator-scripts\") pod \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\" (UID: \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843005 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk4rh\" (UniqueName: \"kubernetes.io/projected/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-kube-api-access-vk4rh\") pod \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\" (UID: \"5971a24c-30d3-4e3e-a649-6d6f637dbfd5\") " Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843622 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkr5r\" (UniqueName: \"kubernetes.io/projected/0f0bf352-90f0-4613-b6c8-9206ddbc4235-kube-api-access-bkr5r\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843653 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f0bf352-90f0-4613-b6c8-9206ddbc4235-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843671 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hh62\" (UniqueName: \"kubernetes.io/projected/7e555442-e24a-467c-8ab3-426856e59e43-kube-api-access-4hh62\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843687 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843703 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843721 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e555442-e24a-467c-8ab3-426856e59e43-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843736 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnprm\" (UniqueName: \"kubernetes.io/projected/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-kube-api-access-tnprm\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843752 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdc2s\" (UniqueName: \"kubernetes.io/projected/ad7fc552-1c39-4c32-a9ea-7df1556ce73f-kube-api-access-bdc2s\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843769 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4e6768f-7cd7-4a37-ab83-59b8fa562c11-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.843784 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr9z8\" (UniqueName: \"kubernetes.io/projected/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051-kube-api-access-hr9z8\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.844922 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5971a24c-30d3-4e3e-a649-6d6f637dbfd5" (UID: "5971a24c-30d3-4e3e-a649-6d6f637dbfd5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.846900 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-kube-api-access-vk4rh" (OuterVolumeSpecName: "kube-api-access-vk4rh") pod "5971a24c-30d3-4e3e-a649-6d6f637dbfd5" (UID: "5971a24c-30d3-4e3e-a649-6d6f637dbfd5"). InnerVolumeSpecName "kube-api-access-vk4rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.901232 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b948-account-create-update-q4k64" event={"ID":"0f0bf352-90f0-4613-b6c8-9206ddbc4235","Type":"ContainerDied","Data":"8051930d46bbd871ebd5641a53a72f1fde6dc3d95fc45deef65dc9261ab97164"} Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.901276 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8051930d46bbd871ebd5641a53a72f1fde6dc3d95fc45deef65dc9261ab97164" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.901346 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b948-account-create-update-q4k64" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.911941 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hgsfq" event={"ID":"5971a24c-30d3-4e3e-a649-6d6f637dbfd5","Type":"ContainerDied","Data":"461d47bb89ec1a340dba420aa36489f917e54b06014617f300ea3aca3fa8dbaa"} Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.911982 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="461d47bb89ec1a340dba420aa36489f917e54b06014617f300ea3aca3fa8dbaa" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.912038 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hgsfq" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.914434 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jbrwj" event={"ID":"7e555442-e24a-467c-8ab3-426856e59e43","Type":"ContainerDied","Data":"44ad644e70e13dc4163443a79033a6da63890cae5e3f2a870851b8883d4d68dd"} Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.914484 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44ad644e70e13dc4163443a79033a6da63890cae5e3f2a870851b8883d4d68dd" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.914540 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jbrwj" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.925992 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-88aa-account-create-update-ndhz5" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.925992 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-88aa-account-create-update-ndhz5" event={"ID":"ad7fc552-1c39-4c32-a9ea-7df1556ce73f","Type":"ContainerDied","Data":"444a13afdc03060fccba0589ed5878a47735e2e018edc0ec8397774edbb13c39"} Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.926028 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="444a13afdc03060fccba0589ed5878a47735e2e018edc0ec8397774edbb13c39" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.928645 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-h5vn4" event={"ID":"3e62dcdb-516f-406c-a6dd-d780f6c250bc","Type":"ContainerStarted","Data":"4dec43cbd7de72379ea3fe3da06372a55dceb25a47b97cfb6a9f05613c52507e"} Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.931745 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nz7sg" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.932241 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nz7sg" event={"ID":"a4e6768f-7cd7-4a37-ab83-59b8fa562c11","Type":"ContainerDied","Data":"6cef8d19c2087835afce157b669b8c6bf91b45db02614c74498f5ee69e0ac628"} Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.932279 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cef8d19c2087835afce157b669b8c6bf91b45db02614c74498f5ee69e0ac628" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.935379 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-953f-account-create-update-595f4" event={"ID":"e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051","Type":"ContainerDied","Data":"0512771fbd6232f3e7263c742b8d671f8c340a401fd100eff0d41be7af660582"} Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.935411 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0512771fbd6232f3e7263c742b8d671f8c340a401fd100eff0d41be7af660582" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.935476 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-953f-account-create-update-595f4" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.948965 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk4rh\" (UniqueName: \"kubernetes.io/projected/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-kube-api-access-vk4rh\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.949170 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5971a24c-30d3-4e3e-a649-6d6f637dbfd5-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:50 crc kubenswrapper[4756]: I0203 09:27:50.956412 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-h5vn4" podStartSLOduration=1.854536145 podStartE2EDuration="9.956395592s" podCreationTimestamp="2026-02-03 09:27:41 +0000 UTC" firstStartedPulling="2026-02-03 09:27:42.265653662 +0000 UTC m=+1053.416121037" lastFinishedPulling="2026-02-03 09:27:50.367513109 +0000 UTC m=+1061.517980484" observedRunningTime="2026-02-03 09:27:50.948831891 +0000 UTC m=+1062.099299286" watchObservedRunningTime="2026-02-03 09:27:50.956395592 +0000 UTC m=+1062.106862967" Feb 03 09:27:54 crc kubenswrapper[4756]: I0203 09:27:54.584607 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:27:54 crc kubenswrapper[4756]: I0203 09:27:54.638528 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dds6r"] Feb 03 09:27:54 crc kubenswrapper[4756]: I0203 09:27:54.638776 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" podUID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" containerName="dnsmasq-dns" containerID="cri-o://d9b900cb37e1258bc3cb26422229a731e02bef084ec5edcfe634acf9aca0b526" gracePeriod=10 Feb 03 09:27:54 crc kubenswrapper[4756]: I0203 09:27:54.972490 4756 generic.go:334] "Generic (PLEG): container finished" podID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" containerID="d9b900cb37e1258bc3cb26422229a731e02bef084ec5edcfe634acf9aca0b526" exitCode=0 Feb 03 09:27:54 crc kubenswrapper[4756]: I0203 09:27:54.972538 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" event={"ID":"1400dee3-79cd-4c19-b349-59c3d1b0bc5c","Type":"ContainerDied","Data":"d9b900cb37e1258bc3cb26422229a731e02bef084ec5edcfe634acf9aca0b526"} Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.142323 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.237348 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-dns-svc\") pod \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.237418 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pls4\" (UniqueName: \"kubernetes.io/projected/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-kube-api-access-8pls4\") pod \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.237506 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-config\") pod \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.237541 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-sb\") pod \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.237632 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-nb\") pod \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\" (UID: \"1400dee3-79cd-4c19-b349-59c3d1b0bc5c\") " Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.254216 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-kube-api-access-8pls4" (OuterVolumeSpecName: "kube-api-access-8pls4") pod "1400dee3-79cd-4c19-b349-59c3d1b0bc5c" (UID: "1400dee3-79cd-4c19-b349-59c3d1b0bc5c"). InnerVolumeSpecName "kube-api-access-8pls4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.296267 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-config" (OuterVolumeSpecName: "config") pod "1400dee3-79cd-4c19-b349-59c3d1b0bc5c" (UID: "1400dee3-79cd-4c19-b349-59c3d1b0bc5c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.301743 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1400dee3-79cd-4c19-b349-59c3d1b0bc5c" (UID: "1400dee3-79cd-4c19-b349-59c3d1b0bc5c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.303768 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1400dee3-79cd-4c19-b349-59c3d1b0bc5c" (UID: "1400dee3-79cd-4c19-b349-59c3d1b0bc5c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.311145 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1400dee3-79cd-4c19-b349-59c3d1b0bc5c" (UID: "1400dee3-79cd-4c19-b349-59c3d1b0bc5c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.340218 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.340268 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.340283 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pls4\" (UniqueName: \"kubernetes.io/projected/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-kube-api-access-8pls4\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.340295 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.340304 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1400dee3-79cd-4c19-b349-59c3d1b0bc5c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.981474 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" event={"ID":"1400dee3-79cd-4c19-b349-59c3d1b0bc5c","Type":"ContainerDied","Data":"fa3c255e624fd63e2d605b897ddd977597db84881449402008a48d2ea3944e29"} Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.981491 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-dds6r" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.982429 4756 scope.go:117] "RemoveContainer" containerID="d9b900cb37e1258bc3cb26422229a731e02bef084ec5edcfe634acf9aca0b526" Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.982959 4756 generic.go:334] "Generic (PLEG): container finished" podID="3e62dcdb-516f-406c-a6dd-d780f6c250bc" containerID="4dec43cbd7de72379ea3fe3da06372a55dceb25a47b97cfb6a9f05613c52507e" exitCode=0 Feb 03 09:27:55 crc kubenswrapper[4756]: I0203 09:27:55.983029 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-h5vn4" event={"ID":"3e62dcdb-516f-406c-a6dd-d780f6c250bc","Type":"ContainerDied","Data":"4dec43cbd7de72379ea3fe3da06372a55dceb25a47b97cfb6a9f05613c52507e"} Feb 03 09:27:56 crc kubenswrapper[4756]: I0203 09:27:56.008966 4756 scope.go:117] "RemoveContainer" containerID="64cd5c06f0373d780a3502feb93099b27653ad1c254c0232dde1169509822c46" Feb 03 09:27:56 crc kubenswrapper[4756]: I0203 09:27:56.018592 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dds6r"] Feb 03 09:27:56 crc kubenswrapper[4756]: I0203 09:27:56.029119 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-dds6r"] Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.289512 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.469532 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-config-data\") pod \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.469590 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-combined-ca-bundle\") pod \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.469788 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lvg5\" (UniqueName: \"kubernetes.io/projected/3e62dcdb-516f-406c-a6dd-d780f6c250bc-kube-api-access-5lvg5\") pod \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\" (UID: \"3e62dcdb-516f-406c-a6dd-d780f6c250bc\") " Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.476924 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e62dcdb-516f-406c-a6dd-d780f6c250bc-kube-api-access-5lvg5" (OuterVolumeSpecName: "kube-api-access-5lvg5") pod "3e62dcdb-516f-406c-a6dd-d780f6c250bc" (UID: "3e62dcdb-516f-406c-a6dd-d780f6c250bc"). InnerVolumeSpecName "kube-api-access-5lvg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.503054 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e62dcdb-516f-406c-a6dd-d780f6c250bc" (UID: "3e62dcdb-516f-406c-a6dd-d780f6c250bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.515205 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-config-data" (OuterVolumeSpecName: "config-data") pod "3e62dcdb-516f-406c-a6dd-d780f6c250bc" (UID: "3e62dcdb-516f-406c-a6dd-d780f6c250bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.571994 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lvg5\" (UniqueName: \"kubernetes.io/projected/3e62dcdb-516f-406c-a6dd-d780f6c250bc-kube-api-access-5lvg5\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.572038 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.572049 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e62dcdb-516f-406c-a6dd-d780f6c250bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:27:57 crc kubenswrapper[4756]: I0203 09:27:57.626829 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" path="/var/lib/kubelet/pods/1400dee3-79cd-4c19-b349-59c3d1b0bc5c/volumes" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.003252 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-h5vn4" event={"ID":"3e62dcdb-516f-406c-a6dd-d780f6c250bc","Type":"ContainerDied","Data":"bb447635f1a9aa81b44e53aec1860c5a1f5669a8dc2bbb4d1660637c2744d25e"} Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.003302 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb447635f1a9aa81b44e53aec1860c5a1f5669a8dc2bbb4d1660637c2744d25e" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.003368 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-h5vn4" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.214113 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-rd5gr"] Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218254 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" containerName="dnsmasq-dns" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218288 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" containerName="dnsmasq-dns" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218305 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" containerName="init" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218315 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" containerName="init" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218326 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218334 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218350 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e555442-e24a-467c-8ab3-426856e59e43" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218358 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e555442-e24a-467c-8ab3-426856e59e43" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218372 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e62dcdb-516f-406c-a6dd-d780f6c250bc" containerName="keystone-db-sync" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218381 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e62dcdb-516f-406c-a6dd-d780f6c250bc" containerName="keystone-db-sync" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218393 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e6768f-7cd7-4a37-ab83-59b8fa562c11" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218400 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e6768f-7cd7-4a37-ab83-59b8fa562c11" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218414 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5971a24c-30d3-4e3e-a649-6d6f637dbfd5" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218422 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5971a24c-30d3-4e3e-a649-6d6f637dbfd5" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218432 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" containerName="dnsmasq-dns" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218439 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" containerName="dnsmasq-dns" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218468 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0bf352-90f0-4613-b6c8-9206ddbc4235" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218477 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0bf352-90f0-4613-b6c8-9206ddbc4235" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218491 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" containerName="init" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218499 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" containerName="init" Feb 03 09:27:58 crc kubenswrapper[4756]: E0203 09:27:58.218508 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7fc552-1c39-4c32-a9ea-7df1556ce73f" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218516 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7fc552-1c39-4c32-a9ea-7df1556ce73f" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218765 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="1400dee3-79cd-4c19-b349-59c3d1b0bc5c" containerName="dnsmasq-dns" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218785 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5971a24c-30d3-4e3e-a649-6d6f637dbfd5" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218798 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218812 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="35741ef2-602f-4fa0-9a37-d430ddc73dfe" containerName="dnsmasq-dns" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218823 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e62dcdb-516f-406c-a6dd-d780f6c250bc" containerName="keystone-db-sync" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218835 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0bf352-90f0-4613-b6c8-9206ddbc4235" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218845 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e6768f-7cd7-4a37-ab83-59b8fa562c11" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218862 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e555442-e24a-467c-8ab3-426856e59e43" containerName="mariadb-database-create" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.218875 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad7fc552-1c39-4c32-a9ea-7df1556ce73f" containerName="mariadb-account-create-update" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.219898 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.235700 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-rd5gr"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.288962 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2lz7z"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.290359 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.293438 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.293716 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fh27s" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.293961 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.297076 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.297333 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.309399 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2lz7z"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.386896 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhlks\" (UniqueName: \"kubernetes.io/projected/35642ebf-dbc7-4af9-8162-e3f853092e5a-kube-api-access-mhlks\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.387074 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.387136 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.387294 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.387391 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.387517 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-config\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.486922 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b76dbdb75-2s7h2"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.488493 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.489840 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-fernet-keys\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.489921 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.489957 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490006 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490044 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490071 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-config-data\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490090 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-credential-keys\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490113 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbhnz\" (UniqueName: \"kubernetes.io/projected/be3a60e9-457f-42df-941c-7ef9c0004939-kube-api-access-pbhnz\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490129 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-scripts\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490155 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-config\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490182 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-combined-ca-bundle\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490208 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhlks\" (UniqueName: \"kubernetes.io/projected/35642ebf-dbc7-4af9-8162-e3f853092e5a-kube-api-access-mhlks\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.490977 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.491010 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.492008 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.492806 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.493010 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-config\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.504723 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.504927 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.505045 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-x5tvw" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.505125 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.533340 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhlks\" (UniqueName: \"kubernetes.io/projected/35642ebf-dbc7-4af9-8162-e3f853092e5a-kube-api-access-mhlks\") pod \"dnsmasq-dns-bbf5cc879-rd5gr\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.555181 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b76dbdb75-2s7h2"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.555525 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591218 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-config-data\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591570 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-scripts\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591618 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-config-data\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591656 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-credential-keys\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591685 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbhnz\" (UniqueName: \"kubernetes.io/projected/be3a60e9-457f-42df-941c-7ef9c0004939-kube-api-access-pbhnz\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591711 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-scripts\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591756 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-combined-ca-bundle\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591842 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-fernet-keys\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591877 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spfw8\" (UniqueName: \"kubernetes.io/projected/6c5d1045-4373-470e-82e5-637eae2b636a-kube-api-access-spfw8\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591911 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c5d1045-4373-470e-82e5-637eae2b636a-logs\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.591929 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c5d1045-4373-470e-82e5-637eae2b636a-horizon-secret-key\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.601072 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-scripts\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.623212 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-config-data\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.623858 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-fernet-keys\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.635458 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbhnz\" (UniqueName: \"kubernetes.io/projected/be3a60e9-457f-42df-941c-7ef9c0004939-kube-api-access-pbhnz\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.647348 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-credential-keys\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.651164 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-4ld7v"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.656391 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-combined-ca-bundle\") pod \"keystone-bootstrap-2lz7z\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.663331 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.665981 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wc6nw" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.666190 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.673193 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.695384 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spfw8\" (UniqueName: \"kubernetes.io/projected/6c5d1045-4373-470e-82e5-637eae2b636a-kube-api-access-spfw8\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.695677 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c5d1045-4373-470e-82e5-637eae2b636a-logs\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.695771 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c5d1045-4373-470e-82e5-637eae2b636a-horizon-secret-key\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.695858 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-config-data\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.695937 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-scripts\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.696791 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-scripts\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.697313 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c5d1045-4373-470e-82e5-637eae2b636a-logs\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.698210 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-config-data\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.699311 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-rd5gr"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.766024 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c5d1045-4373-470e-82e5-637eae2b636a-horizon-secret-key\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.766199 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4ld7v"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.797276 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spfw8\" (UniqueName: \"kubernetes.io/projected/6c5d1045-4373-470e-82e5-637eae2b636a-kube-api-access-spfw8\") pod \"horizon-5b76dbdb75-2s7h2\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.820770 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-cbcxp"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.821181 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7dee8740-0310-450c-88fd-363ce0a26d4b-etc-machine-id\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.821507 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-scripts\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.821550 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq64w\" (UniqueName: \"kubernetes.io/projected/7dee8740-0310-450c-88fd-363ce0a26d4b-kube-api-access-hq64w\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.821611 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-combined-ca-bundle\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.822313 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-db-sync-config-data\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.822375 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.822518 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-config-data\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.824128 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.830317 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8hfm6" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.837634 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.854187 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-7dmqr"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.866972 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7dmqr"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.867106 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.869632 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6ptgf" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.869912 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.870036 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.880364 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-cbcxp"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.911129 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-vxpwc"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.912596 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.916879 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.924622 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4llhn"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.926320 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-db-sync-config-data\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.926386 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-config-data\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.926418 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7dee8740-0310-450c-88fd-363ce0a26d4b-etc-machine-id\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.926491 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.926630 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7dee8740-0310-450c-88fd-363ce0a26d4b-etc-machine-id\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.926506 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-scripts\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.926883 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq64w\" (UniqueName: \"kubernetes.io/projected/7dee8740-0310-450c-88fd-363ce0a26d4b-kube-api-access-hq64w\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.926913 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-combined-ca-bundle\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.934692 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.935506 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jpfb7" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.938519 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-vxpwc"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.952396 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.979348 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-combined-ca-bundle\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.980641 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-config-data\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.995879 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:27:58 crc kubenswrapper[4756]: I0203 09:27:58.998282 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.001951 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-db-sync-config-data\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.002273 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.008599 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.009862 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-scripts\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.020895 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4llhn"] Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.026060 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq64w\" (UniqueName: \"kubernetes.io/projected/7dee8740-0310-450c-88fd-363ce0a26d4b-kube-api-access-hq64w\") pod \"cinder-db-sync-4ld7v\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029254 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6684\" (UniqueName: \"kubernetes.io/projected/79b123a2-b4b7-4233-98e5-538664495935-kube-api-access-b6684\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029298 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029319 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b123a2-b4b7-4233-98e5-538664495935-logs\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029351 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dspqj\" (UniqueName: \"kubernetes.io/projected/f6f14305-2e15-4ccc-ae43-230f1f8daf18-kube-api-access-dspqj\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029378 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029415 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72rkb\" (UniqueName: \"kubernetes.io/projected/e1c29293-41e7-4617-a530-c0f0151d3e55-kube-api-access-72rkb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029434 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-config\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029470 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-combined-ca-bundle\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029497 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-combined-ca-bundle\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029514 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8gzp\" (UniqueName: \"kubernetes.io/projected/1d6582a5-c7dc-4444-924e-9b977a3d4378-kube-api-access-z8gzp\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029531 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-scripts\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029549 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-config-data\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029566 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029598 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-db-sync-config-data\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029619 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-combined-ca-bundle\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029633 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-config\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.029653 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.040305 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.099515 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133318 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133645 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-run-httpd\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133689 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72rkb\" (UniqueName: \"kubernetes.io/projected/e1c29293-41e7-4617-a530-c0f0151d3e55-kube-api-access-72rkb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133724 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-config\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133753 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-combined-ca-bundle\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133781 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133827 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-combined-ca-bundle\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133863 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8gzp\" (UniqueName: \"kubernetes.io/projected/1d6582a5-c7dc-4444-924e-9b977a3d4378-kube-api-access-z8gzp\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133887 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-scripts\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133921 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-config-data\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.133955 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134016 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvlz9\" (UniqueName: \"kubernetes.io/projected/d7957443-ab24-47bf-a76e-0f0b65d9ceca-kube-api-access-kvlz9\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134046 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-db-sync-config-data\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134105 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-combined-ca-bundle\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134136 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-config\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134162 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134192 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134259 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6684\" (UniqueName: \"kubernetes.io/projected/79b123a2-b4b7-4233-98e5-538664495935-kube-api-access-b6684\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134282 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-scripts\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134332 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134359 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b123a2-b4b7-4233-98e5-538664495935-logs\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134453 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dspqj\" (UniqueName: \"kubernetes.io/projected/f6f14305-2e15-4ccc-ae43-230f1f8daf18-kube-api-access-dspqj\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134481 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-config-data\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134520 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-log-httpd\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.134554 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.141315 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.145382 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-combined-ca-bundle\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.150903 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-config-data\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.151705 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.154572 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-db-sync-config-data\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.161838 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-config\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.162530 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.170274 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-combined-ca-bundle\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.170677 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b123a2-b4b7-4233-98e5-538664495935-logs\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.171805 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.172813 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-combined-ca-bundle\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.173251 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-899f4cdfc-zcgmf"] Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.177849 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.189306 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-78t56" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.189634 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.189658 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.189884 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.196926 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.209664 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8gzp\" (UniqueName: \"kubernetes.io/projected/1d6582a5-c7dc-4444-924e-9b977a3d4378-kube-api-access-z8gzp\") pod \"barbican-db-sync-cbcxp\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.211322 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-config\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.216943 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72rkb\" (UniqueName: \"kubernetes.io/projected/e1c29293-41e7-4617-a530-c0f0151d3e55-kube-api-access-72rkb\") pod \"dnsmasq-dns-56df8fb6b7-vxpwc\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.226027 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-899f4cdfc-zcgmf"] Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.237131 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.237371 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.237778 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-config-data\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238106 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvlz9\" (UniqueName: \"kubernetes.io/projected/d7957443-ab24-47bf-a76e-0f0b65d9ceca-kube-api-access-kvlz9\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238211 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238364 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-scripts\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238474 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-config-data\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238552 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-log-httpd\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238652 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmnmg\" (UniqueName: \"kubernetes.io/projected/bfffe83e-c656-48b0-a380-0fe1baab1ad2-kube-api-access-nmnmg\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238738 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-logs\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238818 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-run-httpd\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238886 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.238955 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-scripts\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.239018 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.239096 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.240676 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-scripts\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.241108 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-run-httpd\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.241329 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-log-httpd\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.248036 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6684\" (UniqueName: \"kubernetes.io/projected/79b123a2-b4b7-4233-98e5-538664495935-kube-api-access-b6684\") pod \"placement-db-sync-7dmqr\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.250481 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dspqj\" (UniqueName: \"kubernetes.io/projected/f6f14305-2e15-4ccc-ae43-230f1f8daf18-kube-api-access-dspqj\") pod \"neutron-db-sync-4llhn\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.251419 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.252457 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-config-data\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.252672 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-scripts\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.253423 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.279004 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.281591 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.295375 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvlz9\" (UniqueName: \"kubernetes.io/projected/d7957443-ab24-47bf-a76e-0f0b65d9ceca-kube-api-access-kvlz9\") pod \"ceilometer-0\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.317245 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.321212 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.321424 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.336461 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.337516 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7dmqr" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.340933 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344332 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2232749c-cc9f-4635-a1de-e205d886bd20-logs\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344401 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2232749c-cc9f-4635-a1de-e205d886bd20-horizon-secret-key\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344426 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-scripts\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344550 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcwbv\" (UniqueName: \"kubernetes.io/projected/2232749c-cc9f-4635-a1de-e205d886bd20-kube-api-access-pcwbv\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344670 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmnmg\" (UniqueName: \"kubernetes.io/projected/bfffe83e-c656-48b0-a380-0fe1baab1ad2-kube-api-access-nmnmg\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344695 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-logs\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344749 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344780 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-scripts\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344801 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344902 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.345102 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.345283 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-logs\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.345680 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.344922 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.348581 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-config-data\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.348632 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-config-data\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.357161 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.382771 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.383241 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-scripts\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.383895 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.393957 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4llhn" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.395332 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmnmg\" (UniqueName: \"kubernetes.io/projected/bfffe83e-c656-48b0-a380-0fe1baab1ad2-kube-api-access-nmnmg\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.398188 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-config-data\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.429563 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.450354 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.450672 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.450760 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.450838 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-logs\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.450924 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-config-data\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.451050 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2232749c-cc9f-4635-a1de-e205d886bd20-logs\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.451139 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.451209 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2232749c-cc9f-4635-a1de-e205d886bd20-horizon-secret-key\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.451297 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-scripts\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.451392 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcwbv\" (UniqueName: \"kubernetes.io/projected/2232749c-cc9f-4635-a1de-e205d886bd20-kube-api-access-pcwbv\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.451501 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svshn\" (UniqueName: \"kubernetes.io/projected/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-kube-api-access-svshn\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.451584 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.451657 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.452327 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-config-data\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.452863 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-scripts\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.453136 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2232749c-cc9f-4635-a1de-e205d886bd20-logs\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.471050 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2232749c-cc9f-4635-a1de-e205d886bd20-horizon-secret-key\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.492509 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcwbv\" (UniqueName: \"kubernetes.io/projected/2232749c-cc9f-4635-a1de-e205d886bd20-kube-api-access-pcwbv\") pod \"horizon-899f4cdfc-zcgmf\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.553298 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.554702 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.555032 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svshn\" (UniqueName: \"kubernetes.io/projected/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-kube-api-access-svshn\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.555065 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.555090 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.555148 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.555163 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.555184 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.555198 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-logs\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.555678 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-logs\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.558720 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.558884 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.572874 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.579419 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.579790 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.584714 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svshn\" (UniqueName: \"kubernetes.io/projected/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-kube-api-access-svshn\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.588690 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.602928 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.615812 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.762564 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.763509 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-rd5gr"] Feb 03 09:27:59 crc kubenswrapper[4756]: W0203 09:27:59.828496 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35642ebf_dbc7_4af9_8162_e3f853092e5a.slice/crio-c8010b5c0b0f3a2f4df516587a6e40eb681f9adefb4acc560b9e4e6ab8aa98dd WatchSource:0}: Error finding container c8010b5c0b0f3a2f4df516587a6e40eb681f9adefb4acc560b9e4e6ab8aa98dd: Status 404 returned error can't find the container with id c8010b5c0b0f3a2f4df516587a6e40eb681f9adefb4acc560b9e4e6ab8aa98dd Feb 03 09:27:59 crc kubenswrapper[4756]: I0203 09:27:59.926985 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b76dbdb75-2s7h2"] Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.015689 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.052468 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" event={"ID":"35642ebf-dbc7-4af9-8162-e3f853092e5a","Type":"ContainerStarted","Data":"c8010b5c0b0f3a2f4df516587a6e40eb681f9adefb4acc560b9e4e6ab8aa98dd"} Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.055671 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b76dbdb75-2s7h2" event={"ID":"6c5d1045-4373-470e-82e5-637eae2b636a","Type":"ContainerStarted","Data":"eda838c1210c204df780f2802e008313658808a5cdc22905452088351776cee7"} Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.177360 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2lz7z"] Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.429359 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4ld7v"] Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.436693 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4llhn"] Feb 03 09:28:00 crc kubenswrapper[4756]: W0203 09:28:00.439969 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6f14305_2e15_4ccc_ae43_230f1f8daf18.slice/crio-6a097daca6890a4e41dc9468b6452085c084a0b198457c4bc6796984c2c61a3d WatchSource:0}: Error finding container 6a097daca6890a4e41dc9468b6452085c084a0b198457c4bc6796984c2c61a3d: Status 404 returned error can't find the container with id 6a097daca6890a4e41dc9468b6452085c084a0b198457c4bc6796984c2c61a3d Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.913480 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.933249 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-cbcxp"] Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.947871 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7dmqr"] Feb 03 09:28:00 crc kubenswrapper[4756]: I0203 09:28:00.977155 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:28:01 crc kubenswrapper[4756]: W0203 09:28:01.041018 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2232749c_cc9f_4635_a1de_e205d886bd20.slice/crio-8dcda3dae226b2ece12e1470db2743bdef18b89910089818f3963eaf6df4c914 WatchSource:0}: Error finding container 8dcda3dae226b2ece12e1470db2743bdef18b89910089818f3963eaf6df4c914: Status 404 returned error can't find the container with id 8dcda3dae226b2ece12e1470db2743bdef18b89910089818f3963eaf6df4c914 Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.048042 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-vxpwc"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.069269 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b76dbdb75-2s7h2"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.093658 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-899f4cdfc-zcgmf"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.103330 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bb66dd6bf-zpxb2"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.115811 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.117569 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4llhn" event={"ID":"f6f14305-2e15-4ccc-ae43-230f1f8daf18","Type":"ContainerStarted","Data":"6a097daca6890a4e41dc9468b6452085c084a0b198457c4bc6796984c2c61a3d"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.124968 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-899f4cdfc-zcgmf" event={"ID":"2232749c-cc9f-4635-a1de-e205d886bd20","Type":"ContainerStarted","Data":"8dcda3dae226b2ece12e1470db2743bdef18b89910089818f3963eaf6df4c914"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.136426 4756 generic.go:334] "Generic (PLEG): container finished" podID="35642ebf-dbc7-4af9-8162-e3f853092e5a" containerID="0d32464854ed590b6aaf4c8552a56b02cef64e2947fa1e01c4d8f7e7320395c0" exitCode=0 Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.138129 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bb66dd6bf-zpxb2"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.138220 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" event={"ID":"35642ebf-dbc7-4af9-8162-e3f853092e5a","Type":"ContainerDied","Data":"0d32464854ed590b6aaf4c8552a56b02cef64e2947fa1e01c4d8f7e7320395c0"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.156868 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.167079 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.189992 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2lz7z" event={"ID":"be3a60e9-457f-42df-941c-7ef9c0004939","Type":"ContainerStarted","Data":"84892eacef8d7ac3f04e2ee10739ab647f308ed727c7edb44f4bd3b422ee09fd"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.190082 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2lz7z" event={"ID":"be3a60e9-457f-42df-941c-7ef9c0004939","Type":"ContainerStarted","Data":"c505ccbff1f92cfc6b6cf91a38fe5c24ca5abddb5a15a8e7327edef5fefbcb7b"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.211223 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.213873 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9c7b8f6-aef0-404b-b0df-6f716518aa01-horizon-secret-key\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.213957 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-scripts\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.213990 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-config-data\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.214028 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9c7b8f6-aef0-404b-b0df-6f716518aa01-logs\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.214048 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqdpj\" (UniqueName: \"kubernetes.io/projected/a9c7b8f6-aef0-404b-b0df-6f716518aa01-kube-api-access-wqdpj\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.214031 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerStarted","Data":"d4b89e209e17e69db44a1e044eb3c394ecf8a4a217c0c6b6c1bd4d54716d1838"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.220674 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4ld7v" event={"ID":"7dee8740-0310-450c-88fd-363ce0a26d4b","Type":"ContainerStarted","Data":"0257aa128c7f4b4ec0062d1916da66e456f9e3cd26a33a7a371317d49b3bc992"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.233541 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" event={"ID":"e1c29293-41e7-4617-a530-c0f0151d3e55","Type":"ContainerStarted","Data":"feccabcc1db2e5486baac385e84fc91db24f838b5b58fd67c525233986700f56"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.235942 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cbcxp" event={"ID":"1d6582a5-c7dc-4444-924e-9b977a3d4378","Type":"ContainerStarted","Data":"d194f31f8ea72988ec864a0b3fe3aa2754b8ae85a79e9f08732d678a0bee3f9f"} Feb 03 09:28:01 crc kubenswrapper[4756]: W0203 09:28:01.237655 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfffe83e_c656_48b0_a380_0fe1baab1ad2.slice/crio-5ccd98e55c93b37878c0fdccc20fad4ba2e3ec3fb6658a61181870a7fa26bad1 WatchSource:0}: Error finding container 5ccd98e55c93b37878c0fdccc20fad4ba2e3ec3fb6658a61181870a7fa26bad1: Status 404 returned error can't find the container with id 5ccd98e55c93b37878c0fdccc20fad4ba2e3ec3fb6658a61181870a7fa26bad1 Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.240561 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7dmqr" event={"ID":"79b123a2-b4b7-4233-98e5-538664495935","Type":"ContainerStarted","Data":"1e4c967f95e1f77088b1b0487275a07a80638bd6da30ee1d01e717ea271be80d"} Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.243910 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2lz7z" podStartSLOduration=3.243838195 podStartE2EDuration="3.243838195s" podCreationTimestamp="2026-02-03 09:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:01.234109995 +0000 UTC m=+1072.384577370" watchObservedRunningTime="2026-02-03 09:28:01.243838195 +0000 UTC m=+1072.394305570" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.321527 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-config-data\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.321613 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9c7b8f6-aef0-404b-b0df-6f716518aa01-logs\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.321644 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqdpj\" (UniqueName: \"kubernetes.io/projected/a9c7b8f6-aef0-404b-b0df-6f716518aa01-kube-api-access-wqdpj\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.321736 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9c7b8f6-aef0-404b-b0df-6f716518aa01-horizon-secret-key\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.321846 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-scripts\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.322648 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-scripts\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.325083 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-config-data\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.325208 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9c7b8f6-aef0-404b-b0df-6f716518aa01-logs\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.342943 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9c7b8f6-aef0-404b-b0df-6f716518aa01-horizon-secret-key\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.350077 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqdpj\" (UniqueName: \"kubernetes.io/projected/a9c7b8f6-aef0-404b-b0df-6f716518aa01-kube-api-access-wqdpj\") pod \"horizon-7bb66dd6bf-zpxb2\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.461403 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.534173 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.679193 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-nb\") pod \"35642ebf-dbc7-4af9-8162-e3f853092e5a\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.679264 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-svc\") pod \"35642ebf-dbc7-4af9-8162-e3f853092e5a\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.679368 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-config\") pod \"35642ebf-dbc7-4af9-8162-e3f853092e5a\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.679426 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhlks\" (UniqueName: \"kubernetes.io/projected/35642ebf-dbc7-4af9-8162-e3f853092e5a-kube-api-access-mhlks\") pod \"35642ebf-dbc7-4af9-8162-e3f853092e5a\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.679820 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-swift-storage-0\") pod \"35642ebf-dbc7-4af9-8162-e3f853092e5a\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.679878 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-sb\") pod \"35642ebf-dbc7-4af9-8162-e3f853092e5a\" (UID: \"35642ebf-dbc7-4af9-8162-e3f853092e5a\") " Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.689135 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35642ebf-dbc7-4af9-8162-e3f853092e5a-kube-api-access-mhlks" (OuterVolumeSpecName: "kube-api-access-mhlks") pod "35642ebf-dbc7-4af9-8162-e3f853092e5a" (UID: "35642ebf-dbc7-4af9-8162-e3f853092e5a"). InnerVolumeSpecName "kube-api-access-mhlks". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.707398 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "35642ebf-dbc7-4af9-8162-e3f853092e5a" (UID: "35642ebf-dbc7-4af9-8162-e3f853092e5a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.726585 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35642ebf-dbc7-4af9-8162-e3f853092e5a" (UID: "35642ebf-dbc7-4af9-8162-e3f853092e5a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.727042 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35642ebf-dbc7-4af9-8162-e3f853092e5a" (UID: "35642ebf-dbc7-4af9-8162-e3f853092e5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.746598 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35642ebf-dbc7-4af9-8162-e3f853092e5a" (UID: "35642ebf-dbc7-4af9-8162-e3f853092e5a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.782935 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.782974 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.782987 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhlks\" (UniqueName: \"kubernetes.io/projected/35642ebf-dbc7-4af9-8162-e3f853092e5a-kube-api-access-mhlks\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.783000 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.783010 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.865169 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-config" (OuterVolumeSpecName: "config") pod "35642ebf-dbc7-4af9-8162-e3f853092e5a" (UID: "35642ebf-dbc7-4af9-8162-e3f853092e5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.879186 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:28:01 crc kubenswrapper[4756]: I0203 09:28:01.884373 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35642ebf-dbc7-4af9-8162-e3f853092e5a-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.065959 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bb66dd6bf-zpxb2"] Feb 03 09:28:02 crc kubenswrapper[4756]: W0203 09:28:02.086031 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9c7b8f6_aef0_404b_b0df_6f716518aa01.slice/crio-0c961a97c561677505b4b8e97b5c972c226359347da9f02f465a6247cb25a600 WatchSource:0}: Error finding container 0c961a97c561677505b4b8e97b5c972c226359347da9f02f465a6247cb25a600: Status 404 returned error can't find the container with id 0c961a97c561677505b4b8e97b5c972c226359347da9f02f465a6247cb25a600 Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.309291 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfffe83e-c656-48b0-a380-0fe1baab1ad2","Type":"ContainerStarted","Data":"2747d7c0fc694f6cc7ee5245838087bb254238ac524796145aeff4821aa329e2"} Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.309340 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfffe83e-c656-48b0-a380-0fe1baab1ad2","Type":"ContainerStarted","Data":"5ccd98e55c93b37878c0fdccc20fad4ba2e3ec3fb6658a61181870a7fa26bad1"} Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.313365 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bb66dd6bf-zpxb2" event={"ID":"a9c7b8f6-aef0-404b-b0df-6f716518aa01","Type":"ContainerStarted","Data":"0c961a97c561677505b4b8e97b5c972c226359347da9f02f465a6247cb25a600"} Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.315799 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4llhn" event={"ID":"f6f14305-2e15-4ccc-ae43-230f1f8daf18","Type":"ContainerStarted","Data":"af49ff2cd7d568d6589b3743ca0d88d0f53bcb4d7a4d0881699757d727ee9212"} Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.321849 4756 generic.go:334] "Generic (PLEG): container finished" podID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerID="3ef7348df72bc7a80d4d77cceaefba5c8a8a103f36fb03276e5b64034b65b01a" exitCode=0 Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.321929 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" event={"ID":"e1c29293-41e7-4617-a530-c0f0151d3e55","Type":"ContainerDied","Data":"3ef7348df72bc7a80d4d77cceaefba5c8a8a103f36fb03276e5b64034b65b01a"} Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.327188 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" event={"ID":"35642ebf-dbc7-4af9-8162-e3f853092e5a","Type":"ContainerDied","Data":"c8010b5c0b0f3a2f4df516587a6e40eb681f9adefb4acc560b9e4e6ab8aa98dd"} Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.327233 4756 scope.go:117] "RemoveContainer" containerID="0d32464854ed590b6aaf4c8552a56b02cef64e2947fa1e01c4d8f7e7320395c0" Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.327354 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-rd5gr" Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.338672 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc","Type":"ContainerStarted","Data":"0193fff322a6360b1c3f8bb7a50b92f7d25e304aa45fe73f19392bfa50ff9150"} Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.347867 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4llhn" podStartSLOduration=4.345783248 podStartE2EDuration="4.345783248s" podCreationTimestamp="2026-02-03 09:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:02.33387112 +0000 UTC m=+1073.484338495" watchObservedRunningTime="2026-02-03 09:28:02.345783248 +0000 UTC m=+1073.496250623" Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.420424 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-rd5gr"] Feb 03 09:28:02 crc kubenswrapper[4756]: I0203 09:28:02.443077 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-rd5gr"] Feb 03 09:28:03 crc kubenswrapper[4756]: I0203 09:28:03.365253 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc","Type":"ContainerStarted","Data":"0ee340c116544071b6d93b654a338f5b43849949cec087b3cb8e0e373e0fd30f"} Feb 03 09:28:03 crc kubenswrapper[4756]: I0203 09:28:03.369965 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" event={"ID":"e1c29293-41e7-4617-a530-c0f0151d3e55","Type":"ContainerStarted","Data":"b5f03b1082932810dc6dbce64c10649f12dca9ed05c19c3646acf5a50d9c7e15"} Feb 03 09:28:03 crc kubenswrapper[4756]: I0203 09:28:03.370073 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:28:03 crc kubenswrapper[4756]: I0203 09:28:03.390606 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" podStartSLOduration=5.390585995 podStartE2EDuration="5.390585995s" podCreationTimestamp="2026-02-03 09:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:03.389140669 +0000 UTC m=+1074.539608044" watchObservedRunningTime="2026-02-03 09:28:03.390585995 +0000 UTC m=+1074.541053370" Feb 03 09:28:03 crc kubenswrapper[4756]: I0203 09:28:03.629571 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35642ebf-dbc7-4af9-8162-e3f853092e5a" path="/var/lib/kubelet/pods/35642ebf-dbc7-4af9-8162-e3f853092e5a/volumes" Feb 03 09:28:04 crc kubenswrapper[4756]: I0203 09:28:04.406017 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfffe83e-c656-48b0-a380-0fe1baab1ad2","Type":"ContainerStarted","Data":"937873fd75b00120879e820bd657464a1738abaf3ebedf186638b5ffae5a8b20"} Feb 03 09:28:04 crc kubenswrapper[4756]: I0203 09:28:04.406396 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerName="glance-log" containerID="cri-o://2747d7c0fc694f6cc7ee5245838087bb254238ac524796145aeff4821aa329e2" gracePeriod=30 Feb 03 09:28:04 crc kubenswrapper[4756]: I0203 09:28:04.406956 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerName="glance-httpd" containerID="cri-o://937873fd75b00120879e820bd657464a1738abaf3ebedf186638b5ffae5a8b20" gracePeriod=30 Feb 03 09:28:04 crc kubenswrapper[4756]: I0203 09:28:04.415459 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc","Type":"ContainerStarted","Data":"435f5ffce5b393162bbedfdf02946f97a655623ff0c8a291616a99640de69763"} Feb 03 09:28:04 crc kubenswrapper[4756]: I0203 09:28:04.415522 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerName="glance-log" containerID="cri-o://0ee340c116544071b6d93b654a338f5b43849949cec087b3cb8e0e373e0fd30f" gracePeriod=30 Feb 03 09:28:04 crc kubenswrapper[4756]: I0203 09:28:04.415750 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerName="glance-httpd" containerID="cri-o://435f5ffce5b393162bbedfdf02946f97a655623ff0c8a291616a99640de69763" gracePeriod=30 Feb 03 09:28:04 crc kubenswrapper[4756]: I0203 09:28:04.437025 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.437008202 podStartE2EDuration="6.437008202s" podCreationTimestamp="2026-02-03 09:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:04.432642014 +0000 UTC m=+1075.583109389" watchObservedRunningTime="2026-02-03 09:28:04.437008202 +0000 UTC m=+1075.587475577" Feb 03 09:28:04 crc kubenswrapper[4756]: I0203 09:28:04.477886 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.477857932 podStartE2EDuration="5.477857932s" podCreationTimestamp="2026-02-03 09:27:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:04.475434895 +0000 UTC m=+1075.625902270" watchObservedRunningTime="2026-02-03 09:28:04.477857932 +0000 UTC m=+1075.628325307" Feb 03 09:28:05 crc kubenswrapper[4756]: I0203 09:28:05.427253 4756 generic.go:334] "Generic (PLEG): container finished" podID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerID="435f5ffce5b393162bbedfdf02946f97a655623ff0c8a291616a99640de69763" exitCode=0 Feb 03 09:28:05 crc kubenswrapper[4756]: I0203 09:28:05.427563 4756 generic.go:334] "Generic (PLEG): container finished" podID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerID="0ee340c116544071b6d93b654a338f5b43849949cec087b3cb8e0e373e0fd30f" exitCode=143 Feb 03 09:28:05 crc kubenswrapper[4756]: I0203 09:28:05.427601 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc","Type":"ContainerDied","Data":"435f5ffce5b393162bbedfdf02946f97a655623ff0c8a291616a99640de69763"} Feb 03 09:28:05 crc kubenswrapper[4756]: I0203 09:28:05.427627 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc","Type":"ContainerDied","Data":"0ee340c116544071b6d93b654a338f5b43849949cec087b3cb8e0e373e0fd30f"} Feb 03 09:28:05 crc kubenswrapper[4756]: I0203 09:28:05.430223 4756 generic.go:334] "Generic (PLEG): container finished" podID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerID="937873fd75b00120879e820bd657464a1738abaf3ebedf186638b5ffae5a8b20" exitCode=0 Feb 03 09:28:05 crc kubenswrapper[4756]: I0203 09:28:05.430242 4756 generic.go:334] "Generic (PLEG): container finished" podID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerID="2747d7c0fc694f6cc7ee5245838087bb254238ac524796145aeff4821aa329e2" exitCode=143 Feb 03 09:28:05 crc kubenswrapper[4756]: I0203 09:28:05.430259 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfffe83e-c656-48b0-a380-0fe1baab1ad2","Type":"ContainerDied","Data":"937873fd75b00120879e820bd657464a1738abaf3ebedf186638b5ffae5a8b20"} Feb 03 09:28:05 crc kubenswrapper[4756]: I0203 09:28:05.430276 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfffe83e-c656-48b0-a380-0fe1baab1ad2","Type":"ContainerDied","Data":"2747d7c0fc694f6cc7ee5245838087bb254238ac524796145aeff4821aa329e2"} Feb 03 09:28:06 crc kubenswrapper[4756]: I0203 09:28:06.443028 4756 generic.go:334] "Generic (PLEG): container finished" podID="be3a60e9-457f-42df-941c-7ef9c0004939" containerID="84892eacef8d7ac3f04e2ee10739ab647f308ed727c7edb44f4bd3b422ee09fd" exitCode=0 Feb 03 09:28:06 crc kubenswrapper[4756]: I0203 09:28:06.443139 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2lz7z" event={"ID":"be3a60e9-457f-42df-941c-7ef9c0004939","Type":"ContainerDied","Data":"84892eacef8d7ac3f04e2ee10739ab647f308ed727c7edb44f4bd3b422ee09fd"} Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.783894 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-899f4cdfc-zcgmf"] Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.838682 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-546db499dd-c5f8j"] Feb 03 09:28:07 crc kubenswrapper[4756]: E0203 09:28:07.839086 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35642ebf-dbc7-4af9-8162-e3f853092e5a" containerName="init" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.839104 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="35642ebf-dbc7-4af9-8162-e3f853092e5a" containerName="init" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.839332 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="35642ebf-dbc7-4af9-8162-e3f853092e5a" containerName="init" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.840190 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.844471 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.849675 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-546db499dd-c5f8j"] Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.915801 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-combined-ca-bundle\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.915856 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-scripts\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.919388 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-tls-certs\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.919452 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-config-data\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.919492 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s442\" (UniqueName: \"kubernetes.io/projected/e45bd2fd-140c-409a-9592-252f4723354e-kube-api-access-5s442\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.919581 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-secret-key\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.919627 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45bd2fd-140c-409a-9592-252f4723354e-logs\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.919962 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bb66dd6bf-zpxb2"] Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.964170 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d95f7b7bb-r6fd8"] Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.966023 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:07 crc kubenswrapper[4756]: I0203 09:28:07.980539 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d95f7b7bb-r6fd8"] Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.021839 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-scripts\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.021961 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-tls-certs\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.021991 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-config-data\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.022023 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s442\" (UniqueName: \"kubernetes.io/projected/e45bd2fd-140c-409a-9592-252f4723354e-kube-api-access-5s442\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.022066 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-secret-key\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.022087 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45bd2fd-140c-409a-9592-252f4723354e-logs\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.022126 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-combined-ca-bundle\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.027850 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-scripts\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.033099 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-config-data\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.034003 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-secret-key\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.034503 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-combined-ca-bundle\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.052899 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45bd2fd-140c-409a-9592-252f4723354e-logs\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.053251 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-tls-certs\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.053527 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s442\" (UniqueName: \"kubernetes.io/projected/e45bd2fd-140c-409a-9592-252f4723354e-kube-api-access-5s442\") pod \"horizon-546db499dd-c5f8j\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.124905 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c34fbdee-8998-4e2d-bc98-c5c73ff35160-logs\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.125471 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-combined-ca-bundle\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.125582 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c34fbdee-8998-4e2d-bc98-c5c73ff35160-scripts\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.125623 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb6n4\" (UniqueName: \"kubernetes.io/projected/c34fbdee-8998-4e2d-bc98-c5c73ff35160-kube-api-access-rb6n4\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.125649 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-horizon-secret-key\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.125672 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-horizon-tls-certs\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.125748 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c34fbdee-8998-4e2d-bc98-c5c73ff35160-config-data\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.163771 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.227367 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-combined-ca-bundle\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.227486 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c34fbdee-8998-4e2d-bc98-c5c73ff35160-scripts\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.227521 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb6n4\" (UniqueName: \"kubernetes.io/projected/c34fbdee-8998-4e2d-bc98-c5c73ff35160-kube-api-access-rb6n4\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.227540 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-horizon-secret-key\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.227558 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-horizon-tls-certs\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.227662 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c34fbdee-8998-4e2d-bc98-c5c73ff35160-config-data\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.227719 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c34fbdee-8998-4e2d-bc98-c5c73ff35160-logs\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.228530 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c34fbdee-8998-4e2d-bc98-c5c73ff35160-logs\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.228591 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c34fbdee-8998-4e2d-bc98-c5c73ff35160-scripts\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.230120 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c34fbdee-8998-4e2d-bc98-c5c73ff35160-config-data\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.232005 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-horizon-secret-key\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.234479 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-combined-ca-bundle\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.235225 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c34fbdee-8998-4e2d-bc98-c5c73ff35160-horizon-tls-certs\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.248577 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb6n4\" (UniqueName: \"kubernetes.io/projected/c34fbdee-8998-4e2d-bc98-c5c73ff35160-kube-api-access-rb6n4\") pod \"horizon-6d95f7b7bb-r6fd8\" (UID: \"c34fbdee-8998-4e2d-bc98-c5c73ff35160\") " pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:08 crc kubenswrapper[4756]: I0203 09:28:08.289982 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:09 crc kubenswrapper[4756]: I0203 09:28:09.360574 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:28:09 crc kubenswrapper[4756]: I0203 09:28:09.416531 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-vmwts"] Feb 03 09:28:09 crc kubenswrapper[4756]: I0203 09:28:09.416743 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="dnsmasq-dns" containerID="cri-o://10211dd17ec07922acf6c530bd708fb3d69f765c93b273fe332609b84fcb8994" gracePeriod=10 Feb 03 09:28:09 crc kubenswrapper[4756]: I0203 09:28:09.584904 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Feb 03 09:28:10 crc kubenswrapper[4756]: I0203 09:28:10.489419 4756 generic.go:334] "Generic (PLEG): container finished" podID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerID="10211dd17ec07922acf6c530bd708fb3d69f765c93b273fe332609b84fcb8994" exitCode=0 Feb 03 09:28:10 crc kubenswrapper[4756]: I0203 09:28:10.489487 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" event={"ID":"3c1ee136-d3b1-4d03-8313-81dd9c0f882b","Type":"ContainerDied","Data":"10211dd17ec07922acf6c530bd708fb3d69f765c93b273fe332609b84fcb8994"} Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.234947 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.415894 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-credential-keys\") pod \"be3a60e9-457f-42df-941c-7ef9c0004939\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.416012 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-scripts\") pod \"be3a60e9-457f-42df-941c-7ef9c0004939\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.416040 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-fernet-keys\") pod \"be3a60e9-457f-42df-941c-7ef9c0004939\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.416129 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-combined-ca-bundle\") pod \"be3a60e9-457f-42df-941c-7ef9c0004939\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.416165 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-config-data\") pod \"be3a60e9-457f-42df-941c-7ef9c0004939\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.416912 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbhnz\" (UniqueName: \"kubernetes.io/projected/be3a60e9-457f-42df-941c-7ef9c0004939-kube-api-access-pbhnz\") pod \"be3a60e9-457f-42df-941c-7ef9c0004939\" (UID: \"be3a60e9-457f-42df-941c-7ef9c0004939\") " Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.434627 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "be3a60e9-457f-42df-941c-7ef9c0004939" (UID: "be3a60e9-457f-42df-941c-7ef9c0004939"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.434652 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "be3a60e9-457f-42df-941c-7ef9c0004939" (UID: "be3a60e9-457f-42df-941c-7ef9c0004939"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.435167 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be3a60e9-457f-42df-941c-7ef9c0004939-kube-api-access-pbhnz" (OuterVolumeSpecName: "kube-api-access-pbhnz") pod "be3a60e9-457f-42df-941c-7ef9c0004939" (UID: "be3a60e9-457f-42df-941c-7ef9c0004939"). InnerVolumeSpecName "kube-api-access-pbhnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.437435 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-scripts" (OuterVolumeSpecName: "scripts") pod "be3a60e9-457f-42df-941c-7ef9c0004939" (UID: "be3a60e9-457f-42df-941c-7ef9c0004939"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.445372 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-config-data" (OuterVolumeSpecName: "config-data") pod "be3a60e9-457f-42df-941c-7ef9c0004939" (UID: "be3a60e9-457f-42df-941c-7ef9c0004939"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.448022 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be3a60e9-457f-42df-941c-7ef9c0004939" (UID: "be3a60e9-457f-42df-941c-7ef9c0004939"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.501281 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2lz7z" event={"ID":"be3a60e9-457f-42df-941c-7ef9c0004939","Type":"ContainerDied","Data":"c505ccbff1f92cfc6b6cf91a38fe5c24ca5abddb5a15a8e7327edef5fefbcb7b"} Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.501322 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c505ccbff1f92cfc6b6cf91a38fe5c24ca5abddb5a15a8e7327edef5fefbcb7b" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.501375 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2lz7z" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.519168 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.519196 4756 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.519206 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.519215 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.519224 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbhnz\" (UniqueName: \"kubernetes.io/projected/be3a60e9-457f-42df-941c-7ef9c0004939-kube-api-access-pbhnz\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:11 crc kubenswrapper[4756]: I0203 09:28:11.519232 4756 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/be3a60e9-457f-42df-941c-7ef9c0004939-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.324097 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2lz7z"] Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.331362 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2lz7z"] Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.415949 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-x9wrg"] Feb 03 09:28:12 crc kubenswrapper[4756]: E0203 09:28:12.416483 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be3a60e9-457f-42df-941c-7ef9c0004939" containerName="keystone-bootstrap" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.416506 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="be3a60e9-457f-42df-941c-7ef9c0004939" containerName="keystone-bootstrap" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.416763 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="be3a60e9-457f-42df-941c-7ef9c0004939" containerName="keystone-bootstrap" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.417507 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.420036 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fh27s" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.420094 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.420241 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.420742 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.421364 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.428172 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-x9wrg"] Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.536305 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-fernet-keys\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.536382 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-combined-ca-bundle\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.536554 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77kcm\" (UniqueName: \"kubernetes.io/projected/45100e47-7ef9-499c-b831-ea2994af9403-kube-api-access-77kcm\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.536586 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-credential-keys\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.536671 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-config-data\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.536736 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-scripts\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.639136 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-config-data\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.639221 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-scripts\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.639345 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-fernet-keys\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.639382 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-combined-ca-bundle\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.639501 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77kcm\" (UniqueName: \"kubernetes.io/projected/45100e47-7ef9-499c-b831-ea2994af9403-kube-api-access-77kcm\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.639549 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-credential-keys\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.646758 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-combined-ca-bundle\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.650679 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-config-data\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.660495 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-scripts\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.662409 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-credential-keys\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.662516 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77kcm\" (UniqueName: \"kubernetes.io/projected/45100e47-7ef9-499c-b831-ea2994af9403-kube-api-access-77kcm\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.662698 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-fernet-keys\") pod \"keystone-bootstrap-x9wrg\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:12 crc kubenswrapper[4756]: I0203 09:28:12.776020 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:13 crc kubenswrapper[4756]: I0203 09:28:13.566607 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:28:13 crc kubenswrapper[4756]: I0203 09:28:13.566669 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:28:13 crc kubenswrapper[4756]: I0203 09:28:13.625790 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be3a60e9-457f-42df-941c-7ef9c0004939" path="/var/lib/kubelet/pods/be3a60e9-457f-42df-941c-7ef9c0004939/volumes" Feb 03 09:28:14 crc kubenswrapper[4756]: I0203 09:28:14.584032 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Feb 03 09:28:15 crc kubenswrapper[4756]: E0203 09:28:15.960296 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 03 09:28:15 crc kubenswrapper[4756]: E0203 09:28:15.960813 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b7hf6h66dh648h5b6h646h545h7bh67chdbh59ch697h594hb6h656hffhdchfdhf7hb8h56fh66fhdbh5ch8ch59ch7dh54dh65dh55fhf6h645q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pcwbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-899f4cdfc-zcgmf_openstack(2232749c-cc9f-4635-a1de-e205d886bd20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:28:15 crc kubenswrapper[4756]: E0203 09:28:15.962774 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-899f4cdfc-zcgmf" podUID="2232749c-cc9f-4635-a1de-e205d886bd20" Feb 03 09:28:15 crc kubenswrapper[4756]: E0203 09:28:15.990506 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 03 09:28:15 crc kubenswrapper[4756]: E0203 09:28:15.990691 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bfh57dh586h5cdhdbh655h644h9h687h6h544h5dh66h545h68h59dh54fh5f6hdhd9h68dh6h684h7bh9h59chbdhcdh54h5d6h85h95q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-spfw8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5b76dbdb75-2s7h2_openstack(6c5d1045-4373-470e-82e5-637eae2b636a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:28:15 crc kubenswrapper[4756]: E0203 09:28:15.993408 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5b76dbdb75-2s7h2" podUID="6c5d1045-4373-470e-82e5-637eae2b636a" Feb 03 09:28:19 crc kubenswrapper[4756]: I0203 09:28:19.583959 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Feb 03 09:28:19 crc kubenswrapper[4756]: I0203 09:28:19.584661 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:28:20 crc kubenswrapper[4756]: I0203 09:28:20.581714 4756 generic.go:334] "Generic (PLEG): container finished" podID="f6f14305-2e15-4ccc-ae43-230f1f8daf18" containerID="af49ff2cd7d568d6589b3743ca0d88d0f53bcb4d7a4d0881699757d727ee9212" exitCode=0 Feb 03 09:28:20 crc kubenswrapper[4756]: I0203 09:28:20.581781 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4llhn" event={"ID":"f6f14305-2e15-4ccc-ae43-230f1f8daf18","Type":"ContainerDied","Data":"af49ff2cd7d568d6589b3743ca0d88d0f53bcb4d7a4d0881699757d727ee9212"} Feb 03 09:28:24 crc kubenswrapper[4756]: I0203 09:28:24.583833 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.506554 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.512859 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.523060 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.538126 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4llhn" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.543662 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.622157 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.627069 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4llhn" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.628539 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b76dbdb75-2s7h2" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.632273 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-899f4cdfc-zcgmf" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.636383 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674337 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c5d1045-4373-470e-82e5-637eae2b636a-horizon-secret-key\") pod \"6c5d1045-4373-470e-82e5-637eae2b636a\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674394 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2232749c-cc9f-4635-a1de-e205d886bd20-logs\") pod \"2232749c-cc9f-4635-a1de-e205d886bd20\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674422 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-public-tls-certs\") pod \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674456 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-config-data\") pod \"6c5d1045-4373-470e-82e5-637eae2b636a\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674502 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-logs\") pod \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674576 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-httpd-run\") pod \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674601 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-scripts\") pod \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674634 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2232749c-cc9f-4635-a1de-e205d886bd20-horizon-secret-key\") pod \"2232749c-cc9f-4635-a1de-e205d886bd20\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674659 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dspqj\" (UniqueName: \"kubernetes.io/projected/f6f14305-2e15-4ccc-ae43-230f1f8daf18-kube-api-access-dspqj\") pod \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674686 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c5d1045-4373-470e-82e5-637eae2b636a-logs\") pod \"6c5d1045-4373-470e-82e5-637eae2b636a\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674703 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674760 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcwbv\" (UniqueName: \"kubernetes.io/projected/2232749c-cc9f-4635-a1de-e205d886bd20-kube-api-access-pcwbv\") pod \"2232749c-cc9f-4635-a1de-e205d886bd20\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674783 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spfw8\" (UniqueName: \"kubernetes.io/projected/6c5d1045-4373-470e-82e5-637eae2b636a-kube-api-access-spfw8\") pod \"6c5d1045-4373-470e-82e5-637eae2b636a\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674843 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-scripts\") pod \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674871 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-config-data\") pod \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674870 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2232749c-cc9f-4635-a1de-e205d886bd20-logs" (OuterVolumeSpecName: "logs") pod "2232749c-cc9f-4635-a1de-e205d886bd20" (UID: "2232749c-cc9f-4635-a1de-e205d886bd20"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674888 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-internal-tls-certs\") pod \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.674970 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-httpd-run\") pod \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.675018 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-combined-ca-bundle\") pod \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.675070 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-scripts\") pod \"6c5d1045-4373-470e-82e5-637eae2b636a\" (UID: \"6c5d1045-4373-470e-82e5-637eae2b636a\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.675562 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bfffe83e-c656-48b0-a380-0fe1baab1ad2" (UID: "bfffe83e-c656-48b0-a380-0fe1baab1ad2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.675649 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5d1045-4373-470e-82e5-637eae2b636a-logs" (OuterVolumeSpecName: "logs") pod "6c5d1045-4373-470e-82e5-637eae2b636a" (UID: "6c5d1045-4373-470e-82e5-637eae2b636a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.675958 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmnmg\" (UniqueName: \"kubernetes.io/projected/bfffe83e-c656-48b0-a380-0fe1baab1ad2-kube-api-access-nmnmg\") pod \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676001 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-combined-ca-bundle\") pod \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.675999 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" (UID: "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676029 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-combined-ca-bundle\") pod \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676056 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svshn\" (UniqueName: \"kubernetes.io/projected/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-kube-api-access-svshn\") pod \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676078 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-config-data\") pod \"2232749c-cc9f-4635-a1de-e205d886bd20\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676129 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-config-data\") pod \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676150 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-logs\") pod \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\" (UID: \"bfffe83e-c656-48b0-a380-0fe1baab1ad2\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676165 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-scripts\") pod \"2232749c-cc9f-4635-a1de-e205d886bd20\" (UID: \"2232749c-cc9f-4635-a1de-e205d886bd20\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676183 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\" (UID: \"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676198 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-config\") pod \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\" (UID: \"f6f14305-2e15-4ccc-ae43-230f1f8daf18\") " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676598 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-config-data" (OuterVolumeSpecName: "config-data") pod "6c5d1045-4373-470e-82e5-637eae2b636a" (UID: "6c5d1045-4373-470e-82e5-637eae2b636a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676858 4756 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676876 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c5d1045-4373-470e-82e5-637eae2b636a-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676885 4756 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676895 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2232749c-cc9f-4635-a1de-e205d886bd20-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.676903 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.677098 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-logs" (OuterVolumeSpecName: "logs") pod "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" (UID: "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.677298 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-scripts" (OuterVolumeSpecName: "scripts") pod "6c5d1045-4373-470e-82e5-637eae2b636a" (UID: "6c5d1045-4373-470e-82e5-637eae2b636a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.677984 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-config-data" (OuterVolumeSpecName: "config-data") pod "2232749c-cc9f-4635-a1de-e205d886bd20" (UID: "2232749c-cc9f-4635-a1de-e205d886bd20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.679016 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-scripts" (OuterVolumeSpecName: "scripts") pod "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" (UID: "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.679495 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2232749c-cc9f-4635-a1de-e205d886bd20-kube-api-access-pcwbv" (OuterVolumeSpecName: "kube-api-access-pcwbv") pod "2232749c-cc9f-4635-a1de-e205d886bd20" (UID: "2232749c-cc9f-4635-a1de-e205d886bd20"). InnerVolumeSpecName "kube-api-access-pcwbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.679790 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f14305-2e15-4ccc-ae43-230f1f8daf18-kube-api-access-dspqj" (OuterVolumeSpecName: "kube-api-access-dspqj") pod "f6f14305-2e15-4ccc-ae43-230f1f8daf18" (UID: "f6f14305-2e15-4ccc-ae43-230f1f8daf18"). InnerVolumeSpecName "kube-api-access-dspqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.679930 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-scripts" (OuterVolumeSpecName: "scripts") pod "2232749c-cc9f-4635-a1de-e205d886bd20" (UID: "2232749c-cc9f-4635-a1de-e205d886bd20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.680025 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "bfffe83e-c656-48b0-a380-0fe1baab1ad2" (UID: "bfffe83e-c656-48b0-a380-0fe1baab1ad2"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.680201 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-logs" (OuterVolumeSpecName: "logs") pod "bfffe83e-c656-48b0-a380-0fe1baab1ad2" (UID: "bfffe83e-c656-48b0-a380-0fe1baab1ad2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.680221 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c5d1045-4373-470e-82e5-637eae2b636a-kube-api-access-spfw8" (OuterVolumeSpecName: "kube-api-access-spfw8") pod "6c5d1045-4373-470e-82e5-637eae2b636a" (UID: "6c5d1045-4373-470e-82e5-637eae2b636a"). InnerVolumeSpecName "kube-api-access-spfw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.683078 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfffe83e-c656-48b0-a380-0fe1baab1ad2-kube-api-access-nmnmg" (OuterVolumeSpecName: "kube-api-access-nmnmg") pod "bfffe83e-c656-48b0-a380-0fe1baab1ad2" (UID: "bfffe83e-c656-48b0-a380-0fe1baab1ad2"). InnerVolumeSpecName "kube-api-access-nmnmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.683471 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" (UID: "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.685175 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5d1045-4373-470e-82e5-637eae2b636a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6c5d1045-4373-470e-82e5-637eae2b636a" (UID: "6c5d1045-4373-470e-82e5-637eae2b636a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.696374 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-scripts" (OuterVolumeSpecName: "scripts") pod "bfffe83e-c656-48b0-a380-0fe1baab1ad2" (UID: "bfffe83e-c656-48b0-a380-0fe1baab1ad2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.707858 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2232749c-cc9f-4635-a1de-e205d886bd20-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2232749c-cc9f-4635-a1de-e205d886bd20" (UID: "2232749c-cc9f-4635-a1de-e205d886bd20"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.707907 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-kube-api-access-svshn" (OuterVolumeSpecName: "kube-api-access-svshn") pod "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" (UID: "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc"). InnerVolumeSpecName "kube-api-access-svshn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.714923 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" (UID: "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.717033 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfffe83e-c656-48b0-a380-0fe1baab1ad2" (UID: "bfffe83e-c656-48b0-a380-0fe1baab1ad2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.720707 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-config" (OuterVolumeSpecName: "config") pod "f6f14305-2e15-4ccc-ae43-230f1f8daf18" (UID: "f6f14305-2e15-4ccc-ae43-230f1f8daf18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.731662 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6f14305-2e15-4ccc-ae43-230f1f8daf18" (UID: "f6f14305-2e15-4ccc-ae43-230f1f8daf18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.737604 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bfffe83e-c656-48b0-a380-0fe1baab1ad2" (UID: "bfffe83e-c656-48b0-a380-0fe1baab1ad2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.738586 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-config-data" (OuterVolumeSpecName: "config-data") pod "bfffe83e-c656-48b0-a380-0fe1baab1ad2" (UID: "bfffe83e-c656-48b0-a380-0fe1baab1ad2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.740563 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-config-data" (OuterVolumeSpecName: "config-data") pod "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" (UID: "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.742897 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" (UID: "2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778338 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778382 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778396 4756 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2232749c-cc9f-4635-a1de-e205d886bd20-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778413 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dspqj\" (UniqueName: \"kubernetes.io/projected/f6f14305-2e15-4ccc-ae43-230f1f8daf18-kube-api-access-dspqj\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778437 4756 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778466 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcwbv\" (UniqueName: \"kubernetes.io/projected/2232749c-cc9f-4635-a1de-e205d886bd20-kube-api-access-pcwbv\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778479 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spfw8\" (UniqueName: \"kubernetes.io/projected/6c5d1045-4373-470e-82e5-637eae2b636a-kube-api-access-spfw8\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778489 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778500 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778511 4756 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778525 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778537 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c5d1045-4373-470e-82e5-637eae2b636a-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778549 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmnmg\" (UniqueName: \"kubernetes.io/projected/bfffe83e-c656-48b0-a380-0fe1baab1ad2-kube-api-access-nmnmg\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778560 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778572 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778584 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svshn\" (UniqueName: \"kubernetes.io/projected/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-kube-api-access-svshn\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778597 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778608 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778619 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfffe83e-c656-48b0-a380-0fe1baab1ad2-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778629 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2232749c-cc9f-4635-a1de-e205d886bd20-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778648 4756 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778662 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6f14305-2e15-4ccc-ae43-230f1f8daf18-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778675 4756 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c5d1045-4373-470e-82e5-637eae2b636a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.778687 4756 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfffe83e-c656-48b0-a380-0fe1baab1ad2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.799125 4756 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.799916 4756 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.851771 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfffe83e-c656-48b0-a380-0fe1baab1ad2","Type":"ContainerDied","Data":"5ccd98e55c93b37878c0fdccc20fad4ba2e3ec3fb6658a61181870a7fa26bad1"} Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.851813 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4llhn" event={"ID":"f6f14305-2e15-4ccc-ae43-230f1f8daf18","Type":"ContainerDied","Data":"6a097daca6890a4e41dc9468b6452085c084a0b198457c4bc6796984c2c61a3d"} Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.851827 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a097daca6890a4e41dc9468b6452085c084a0b198457c4bc6796984c2c61a3d" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.851843 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b76dbdb75-2s7h2" event={"ID":"6c5d1045-4373-470e-82e5-637eae2b636a","Type":"ContainerDied","Data":"eda838c1210c204df780f2802e008313658808a5cdc22905452088351776cee7"} Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.851853 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-899f4cdfc-zcgmf" event={"ID":"2232749c-cc9f-4635-a1de-e205d886bd20","Type":"ContainerDied","Data":"8dcda3dae226b2ece12e1470db2743bdef18b89910089818f3963eaf6df4c914"} Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.851864 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc","Type":"ContainerDied","Data":"0193fff322a6360b1c3f8bb7a50b92f7d25e304aa45fe73f19392bfa50ff9150"} Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.851884 4756 scope.go:117] "RemoveContainer" containerID="937873fd75b00120879e820bd657464a1738abaf3ebedf186638b5ffae5a8b20" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.879874 4756 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.879901 4756 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.965351 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.975052 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.988562 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:28:25 crc kubenswrapper[4756]: E0203 09:28:25.988946 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerName="glance-httpd" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.988958 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerName="glance-httpd" Feb 03 09:28:25 crc kubenswrapper[4756]: E0203 09:28:25.988971 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerName="glance-log" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.988977 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerName="glance-log" Feb 03 09:28:25 crc kubenswrapper[4756]: E0203 09:28:25.988985 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f14305-2e15-4ccc-ae43-230f1f8daf18" containerName="neutron-db-sync" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.988992 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f14305-2e15-4ccc-ae43-230f1f8daf18" containerName="neutron-db-sync" Feb 03 09:28:25 crc kubenswrapper[4756]: E0203 09:28:25.988999 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerName="glance-httpd" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.989005 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerName="glance-httpd" Feb 03 09:28:25 crc kubenswrapper[4756]: E0203 09:28:25.989017 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerName="glance-log" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.989023 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerName="glance-log" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.989193 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerName="glance-log" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.989205 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerName="glance-httpd" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.989218 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" containerName="glance-httpd" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.989227 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f14305-2e15-4ccc-ae43-230f1f8daf18" containerName="neutron-db-sync" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.989240 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" containerName="glance-log" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.990132 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.995660 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.996078 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.996579 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 03 09:28:25 crc kubenswrapper[4756]: I0203 09:28:25.998616 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-78t56" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.022935 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.063231 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b76dbdb75-2s7h2"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.075675 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5b76dbdb75-2s7h2"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.087931 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-scripts\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.088029 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.088081 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.088192 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.088219 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-logs\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.088264 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-config-data\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.088291 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.088333 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8tv6\" (UniqueName: \"kubernetes.io/projected/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-kube-api-access-l8tv6\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.095400 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-899f4cdfc-zcgmf"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.109603 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-899f4cdfc-zcgmf"] Feb 03 09:28:26 crc kubenswrapper[4756]: E0203 09:28:26.136006 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 03 09:28:26 crc kubenswrapper[4756]: E0203 09:28:26.136186 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z8gzp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-cbcxp_openstack(1d6582a5-c7dc-4444-924e-9b977a3d4378): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:28:26 crc kubenswrapper[4756]: E0203 09:28:26.137889 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-cbcxp" podUID="1d6582a5-c7dc-4444-924e-9b977a3d4378" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.144936 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.159240 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.167769 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.169911 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.171932 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.172296 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.176514 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.189942 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-scripts\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.190007 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.190041 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.190117 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.190140 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-logs\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.194926 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.195320 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-logs\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.195885 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.196072 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-scripts\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.200667 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.201565 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-config-data\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.201637 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.201710 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8tv6\" (UniqueName: \"kubernetes.io/projected/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-kube-api-access-l8tv6\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.205331 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.205479 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-config-data\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.219372 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8tv6\" (UniqueName: \"kubernetes.io/projected/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-kube-api-access-l8tv6\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.226491 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.306888 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-logs\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.306971 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.307009 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.307064 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.307138 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnknm\" (UniqueName: \"kubernetes.io/projected/16c4732e-295a-495d-a2af-e10b2c71a10f-kube-api-access-wnknm\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.307166 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.307209 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.307268 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.310883 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.408874 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.408944 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.408987 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-logs\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.409039 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.409078 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.409110 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.409177 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnknm\" (UniqueName: \"kubernetes.io/projected/16c4732e-295a-495d-a2af-e10b2c71a10f-kube-api-access-wnknm\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.409198 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.409205 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.409515 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-logs\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.409810 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.418147 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.418302 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.420154 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.424035 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.428700 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnknm\" (UniqueName: \"kubernetes.io/projected/16c4732e-295a-495d-a2af-e10b2c71a10f-kube-api-access-wnknm\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.434919 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.490279 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:26 crc kubenswrapper[4756]: E0203 09:28:26.645129 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-cbcxp" podUID="1d6582a5-c7dc-4444-924e-9b977a3d4378" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.830981 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-mm4zx"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.835641 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.846861 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-mm4zx"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.921581 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-svc\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.921666 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-config\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.921722 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.921752 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.921902 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.921925 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxtzc\" (UniqueName: \"kubernetes.io/projected/3beb75ec-64c9-413b-9efe-63a05ddb0bab-kube-api-access-nxtzc\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.977245 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7b74899c58-bsxwf"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.981621 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.990580 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.992040 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b74899c58-bsxwf"] Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.992169 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.992221 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 03 09:28:26 crc kubenswrapper[4756]: I0203 09:28:26.992277 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jpfb7" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.023929 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-config\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.024070 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.024150 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.024519 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.024637 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxtzc\" (UniqueName: \"kubernetes.io/projected/3beb75ec-64c9-413b-9efe-63a05ddb0bab-kube-api-access-nxtzc\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.024731 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-svc\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.025358 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-svc\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.026013 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.026182 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.026762 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.027255 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-config\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.057231 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxtzc\" (UniqueName: \"kubernetes.io/projected/3beb75ec-64c9-413b-9efe-63a05ddb0bab-kube-api-access-nxtzc\") pod \"dnsmasq-dns-6b7b667979-mm4zx\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.126595 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-ovndb-tls-certs\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.126727 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds97c\" (UniqueName: \"kubernetes.io/projected/c571905e-5655-45c7-a864-70d02e7a9326-kube-api-access-ds97c\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.126831 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-httpd-config\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.126875 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-config\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.126897 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-combined-ca-bundle\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.173612 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.228471 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-ovndb-tls-certs\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.228566 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds97c\" (UniqueName: \"kubernetes.io/projected/c571905e-5655-45c7-a864-70d02e7a9326-kube-api-access-ds97c\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.228648 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-httpd-config\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.228682 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-config\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.228703 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-combined-ca-bundle\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.238117 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-ovndb-tls-certs\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.238298 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-combined-ca-bundle\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.238528 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-httpd-config\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.238606 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-config\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.250834 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds97c\" (UniqueName: \"kubernetes.io/projected/c571905e-5655-45c7-a864-70d02e7a9326-kube-api-access-ds97c\") pod \"neutron-7b74899c58-bsxwf\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.310620 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.631657 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2232749c-cc9f-4635-a1de-e205d886bd20" path="/var/lib/kubelet/pods/2232749c-cc9f-4635-a1de-e205d886bd20/volumes" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.632148 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc" path="/var/lib/kubelet/pods/2f7b328b-4e50-4ff0-b838-41f4cfc7c1fc/volumes" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.633139 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c5d1045-4373-470e-82e5-637eae2b636a" path="/var/lib/kubelet/pods/6c5d1045-4373-470e-82e5-637eae2b636a/volumes" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.635778 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfffe83e-c656-48b0-a380-0fe1baab1ad2" path="/var/lib/kubelet/pods/bfffe83e-c656-48b0-a380-0fe1baab1ad2/volumes" Feb 03 09:28:27 crc kubenswrapper[4756]: E0203 09:28:27.747263 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 03 09:28:27 crc kubenswrapper[4756]: E0203 09:28:27.747671 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hq64w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-4ld7v_openstack(7dee8740-0310-450c-88fd-363ce0a26d4b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:28:27 crc kubenswrapper[4756]: E0203 09:28:27.753533 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-4ld7v" podUID="7dee8740-0310-450c-88fd-363ce0a26d4b" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.832658 4756 scope.go:117] "RemoveContainer" containerID="2747d7c0fc694f6cc7ee5245838087bb254238ac524796145aeff4821aa329e2" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.910109 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:28:27 crc kubenswrapper[4756]: I0203 09:28:27.975695 4756 scope.go:117] "RemoveContainer" containerID="435f5ffce5b393162bbedfdf02946f97a655623ff0c8a291616a99640de69763" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.043029 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldtbs\" (UniqueName: \"kubernetes.io/projected/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-kube-api-access-ldtbs\") pod \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.043097 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-sb\") pod \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.043183 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-swift-storage-0\") pod \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.043212 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-svc\") pod \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.043244 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-config\") pod \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.043401 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-nb\") pod \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\" (UID: \"3c1ee136-d3b1-4d03-8313-81dd9c0f882b\") " Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.060955 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-kube-api-access-ldtbs" (OuterVolumeSpecName: "kube-api-access-ldtbs") pod "3c1ee136-d3b1-4d03-8313-81dd9c0f882b" (UID: "3c1ee136-d3b1-4d03-8313-81dd9c0f882b"). InnerVolumeSpecName "kube-api-access-ldtbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.077649 4756 scope.go:117] "RemoveContainer" containerID="0ee340c116544071b6d93b654a338f5b43849949cec087b3cb8e0e373e0fd30f" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.146167 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldtbs\" (UniqueName: \"kubernetes.io/projected/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-kube-api-access-ldtbs\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.169298 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3c1ee136-d3b1-4d03-8313-81dd9c0f882b" (UID: "3c1ee136-d3b1-4d03-8313-81dd9c0f882b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.186558 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3c1ee136-d3b1-4d03-8313-81dd9c0f882b" (UID: "3c1ee136-d3b1-4d03-8313-81dd9c0f882b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.192641 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3c1ee136-d3b1-4d03-8313-81dd9c0f882b" (UID: "3c1ee136-d3b1-4d03-8313-81dd9c0f882b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.197221 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-config" (OuterVolumeSpecName: "config") pod "3c1ee136-d3b1-4d03-8313-81dd9c0f882b" (UID: "3c1ee136-d3b1-4d03-8313-81dd9c0f882b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.207959 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3c1ee136-d3b1-4d03-8313-81dd9c0f882b" (UID: "3c1ee136-d3b1-4d03-8313-81dd9c0f882b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.247826 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.247860 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.247875 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.247887 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.247901 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1ee136-d3b1-4d03-8313-81dd9c0f882b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.603515 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d95f7b7bb-r6fd8"] Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.636787 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-546db499dd-c5f8j"] Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.691572 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7dmqr" event={"ID":"79b123a2-b4b7-4233-98e5-538664495935","Type":"ContainerStarted","Data":"0439d59a66669486fb155370afc9837c626a409281f0397983d81cb77366332d"} Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.735920 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-546db499dd-c5f8j" event={"ID":"e45bd2fd-140c-409a-9592-252f4723354e","Type":"ContainerStarted","Data":"b0fa5b1cbc3eeb016c58e31ffcbd8b47631bcf967d79d7e8aba7d7056b011bc2"} Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.741040 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-7dmqr" podStartSLOduration=6.357904916 podStartE2EDuration="30.741024455s" podCreationTimestamp="2026-02-03 09:27:58 +0000 UTC" firstStartedPulling="2026-02-03 09:28:00.987111128 +0000 UTC m=+1072.137578493" lastFinishedPulling="2026-02-03 09:28:25.370230657 +0000 UTC m=+1096.520698032" observedRunningTime="2026-02-03 09:28:28.7138238 +0000 UTC m=+1099.864291175" watchObservedRunningTime="2026-02-03 09:28:28.741024455 +0000 UTC m=+1099.891491830" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.742190 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-x9wrg"] Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.746208 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" event={"ID":"3c1ee136-d3b1-4d03-8313-81dd9c0f882b","Type":"ContainerDied","Data":"b500f8c760b3bc134764b66bea2156424566f48160f8bec8726af5ddf01dd6ed"} Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.746273 4756 scope.go:117] "RemoveContainer" containerID="10211dd17ec07922acf6c530bd708fb3d69f765c93b273fe332609b84fcb8994" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.746329 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-vmwts" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.754037 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerStarted","Data":"72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273"} Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.755792 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d95f7b7bb-r6fd8" event={"ID":"c34fbdee-8998-4e2d-bc98-c5c73ff35160","Type":"ContainerStarted","Data":"2eb9ad39a3d9d2ed0b0ffc6d01b60b41d9e3e91f4f1bc766076e303c7dfa167a"} Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.771077 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bb66dd6bf-zpxb2" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerName="horizon-log" containerID="cri-o://5f5ead35d19b5efc89edc6b946b5e705dec6a56c9da7887aff9733cdc2c84350" gracePeriod=30 Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.771134 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bb66dd6bf-zpxb2" event={"ID":"a9c7b8f6-aef0-404b-b0df-6f716518aa01","Type":"ContainerStarted","Data":"5f5ead35d19b5efc89edc6b946b5e705dec6a56c9da7887aff9733cdc2c84350"} Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.771182 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bb66dd6bf-zpxb2" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerName="horizon" containerID="cri-o://a2e41bbc539623bf400c55db9e7628d7435096b137b557a7fb505514f8eec9c5" gracePeriod=30 Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.777580 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-mm4zx"] Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.788787 4756 scope.go:117] "RemoveContainer" containerID="2aa552b13be2cc0ddb763534153a4d76f01f5fe14d976094d8381330f8593b2d" Feb 03 09:28:28 crc kubenswrapper[4756]: E0203 09:28:28.789016 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-4ld7v" podUID="7dee8740-0310-450c-88fd-363ce0a26d4b" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.831721 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7bb66dd6bf-zpxb2" podStartSLOduration=3.235203483 podStartE2EDuration="28.83170405s" podCreationTimestamp="2026-02-03 09:28:00 +0000 UTC" firstStartedPulling="2026-02-03 09:28:02.091260322 +0000 UTC m=+1073.241727697" lastFinishedPulling="2026-02-03 09:28:27.687760889 +0000 UTC m=+1098.838228264" observedRunningTime="2026-02-03 09:28:28.816604219 +0000 UTC m=+1099.967071604" watchObservedRunningTime="2026-02-03 09:28:28.83170405 +0000 UTC m=+1099.982171425" Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.858148 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-vmwts"] Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.876280 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-vmwts"] Feb 03 09:28:28 crc kubenswrapper[4756]: I0203 09:28:28.933198 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b74899c58-bsxwf"] Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.022430 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.094640 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cdf446f5c-8xnr7"] Feb 03 09:28:29 crc kubenswrapper[4756]: E0203 09:28:29.095567 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="dnsmasq-dns" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.095641 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="dnsmasq-dns" Feb 03 09:28:29 crc kubenswrapper[4756]: E0203 09:28:29.095737 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="init" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.095796 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="init" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.095994 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" containerName="dnsmasq-dns" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.101374 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.110527 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.110720 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.126231 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cdf446f5c-8xnr7"] Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.167808 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-public-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.168068 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-internal-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.168144 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-combined-ca-bundle\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.168233 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-ovndb-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.168349 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-httpd-config\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.168418 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-config\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.168597 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x657q\" (UniqueName: \"kubernetes.io/projected/7fce406f-4aff-48f6-980a-64a65ede0c74-kube-api-access-x657q\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.270063 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x657q\" (UniqueName: \"kubernetes.io/projected/7fce406f-4aff-48f6-980a-64a65ede0c74-kube-api-access-x657q\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.270144 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-public-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.270229 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-internal-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.270255 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-combined-ca-bundle\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.270297 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-ovndb-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.270373 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-httpd-config\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.270402 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-config\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.280424 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-public-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.281430 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-ovndb-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.281642 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-config\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.282059 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-combined-ca-bundle\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.282329 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-httpd-config\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.287555 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x657q\" (UniqueName: \"kubernetes.io/projected/7fce406f-4aff-48f6-980a-64a65ede0c74-kube-api-access-x657q\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.302863 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-internal-tls-certs\") pod \"neutron-6cdf446f5c-8xnr7\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.445043 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.650749 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c1ee136-d3b1-4d03-8313-81dd9c0f882b" path="/var/lib/kubelet/pods/3c1ee136-d3b1-4d03-8313-81dd9c0f882b/volumes" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.821776 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"19ccb5fd-4e2d-4cc8-b776-161687b99eaa","Type":"ContainerStarted","Data":"137b54407711507c81c2f21c0e61b626e51715a3e585a15b9ba37a5327a3fe98"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.824197 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d95f7b7bb-r6fd8" event={"ID":"c34fbdee-8998-4e2d-bc98-c5c73ff35160","Type":"ContainerStarted","Data":"b5629470c3ca538894f6bcd21f75dd18591bd48588dec15e4ab12325a780b3c0"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.824233 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d95f7b7bb-r6fd8" event={"ID":"c34fbdee-8998-4e2d-bc98-c5c73ff35160","Type":"ContainerStarted","Data":"9bd6bdf64760070093e39bbc0c62d21d56176ea7c160e886cc7e8a69ca7ce689"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.839836 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-546db499dd-c5f8j" event={"ID":"e45bd2fd-140c-409a-9592-252f4723354e","Type":"ContainerStarted","Data":"4f5b4620a588f33eab38c31e48cf91bf8ea74628a6ed552028a79834e5626fdb"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.839938 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-546db499dd-c5f8j" event={"ID":"e45bd2fd-140c-409a-9592-252f4723354e","Type":"ContainerStarted","Data":"798d8b64be1bedcbb904e87c52b8e23badc5c7864eb3968e0e041473058e8551"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.857651 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bb66dd6bf-zpxb2" event={"ID":"a9c7b8f6-aef0-404b-b0df-6f716518aa01","Type":"ContainerStarted","Data":"a2e41bbc539623bf400c55db9e7628d7435096b137b557a7fb505514f8eec9c5"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.857717 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d95f7b7bb-r6fd8" podStartSLOduration=22.857678487 podStartE2EDuration="22.857678487s" podCreationTimestamp="2026-02-03 09:28:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:29.842425862 +0000 UTC m=+1100.992893257" watchObservedRunningTime="2026-02-03 09:28:29.857678487 +0000 UTC m=+1101.008145882" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.868903 4756 generic.go:334] "Generic (PLEG): container finished" podID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" containerID="7c2cfc72d439839f4aa8727963f7dd9a27d3f3e373f4d58c6b4ded60402f5db1" exitCode=0 Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.869175 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" event={"ID":"3beb75ec-64c9-413b-9efe-63a05ddb0bab","Type":"ContainerDied","Data":"7c2cfc72d439839f4aa8727963f7dd9a27d3f3e373f4d58c6b4ded60402f5db1"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.869271 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" event={"ID":"3beb75ec-64c9-413b-9efe-63a05ddb0bab","Type":"ContainerStarted","Data":"5489ed7f3a51c5e1ff968562b05333d4b9957c9e8f954ad961b1468a254c34db"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.875893 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-546db499dd-c5f8j" podStartSLOduration=22.875875836 podStartE2EDuration="22.875875836s" podCreationTimestamp="2026-02-03 09:28:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:29.859148624 +0000 UTC m=+1101.009615999" watchObservedRunningTime="2026-02-03 09:28:29.875875836 +0000 UTC m=+1101.026343211" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.931590 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x9wrg" event={"ID":"45100e47-7ef9-499c-b831-ea2994af9403","Type":"ContainerStarted","Data":"66e2f1708ffff8fa5abe63648080d850074bd13d775bd047ec3c99fca2a25235"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.931645 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x9wrg" event={"ID":"45100e47-7ef9-499c-b831-ea2994af9403","Type":"ContainerStarted","Data":"8f1a01d83995e5795d989c6f88a137a96ae6a2487ac2e6e98d1562d1260552ac"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.939085 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b74899c58-bsxwf" event={"ID":"c571905e-5655-45c7-a864-70d02e7a9326","Type":"ContainerStarted","Data":"dedf5fd2bb1e43cfe9147a8997060b222a345753833e11dab8117b21d7572e1d"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.939145 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b74899c58-bsxwf" event={"ID":"c571905e-5655-45c7-a864-70d02e7a9326","Type":"ContainerStarted","Data":"1b0973a2612cd6f68029aef4d5a1476dd1e24253e75900f5476116c0a5b7eca0"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.939168 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b74899c58-bsxwf" event={"ID":"c571905e-5655-45c7-a864-70d02e7a9326","Type":"ContainerStarted","Data":"6bcf1ad35168003f3d1acfc0ecc2eb531505d0eaee0b45e4a7011b050f8968f6"} Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.940632 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:29 crc kubenswrapper[4756]: I0203 09:28:29.970391 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-x9wrg" podStartSLOduration=17.970362801 podStartE2EDuration="17.970362801s" podCreationTimestamp="2026-02-03 09:28:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:29.954228568 +0000 UTC m=+1101.104695943" watchObservedRunningTime="2026-02-03 09:28:29.970362801 +0000 UTC m=+1101.120830176" Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.034105 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7b74899c58-bsxwf" podStartSLOduration=4.034079418 podStartE2EDuration="4.034079418s" podCreationTimestamp="2026-02-03 09:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:29.9879246 +0000 UTC m=+1101.138391975" watchObservedRunningTime="2026-02-03 09:28:30.034079418 +0000 UTC m=+1101.184546793" Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.124714 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:28:30 crc kubenswrapper[4756]: W0203 09:28:30.131372 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16c4732e_295a_495d_a2af_e10b2c71a10f.slice/crio-9edfe39d891b6ce82ffbe58805e9ccfb660934332dce6f97d42ae01582069c23 WatchSource:0}: Error finding container 9edfe39d891b6ce82ffbe58805e9ccfb660934332dce6f97d42ae01582069c23: Status 404 returned error can't find the container with id 9edfe39d891b6ce82ffbe58805e9ccfb660934332dce6f97d42ae01582069c23 Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.210806 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cdf446f5c-8xnr7"] Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.959632 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" event={"ID":"3beb75ec-64c9-413b-9efe-63a05ddb0bab","Type":"ContainerStarted","Data":"2b206a6e16728314c6e93f23fc70347b7c70b457f411ac981811d128d4b7b258"} Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.959710 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.962698 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdf446f5c-8xnr7" event={"ID":"7fce406f-4aff-48f6-980a-64a65ede0c74","Type":"ContainerStarted","Data":"0cb64b5542a53c0cbae8d0bccb77ca32ef6e6ac36f7b885c4b35edac0fdb6a24"} Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.965761 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"19ccb5fd-4e2d-4cc8-b776-161687b99eaa","Type":"ContainerStarted","Data":"910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0"} Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.968273 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16c4732e-295a-495d-a2af-e10b2c71a10f","Type":"ContainerStarted","Data":"9edfe39d891b6ce82ffbe58805e9ccfb660934332dce6f97d42ae01582069c23"} Feb 03 09:28:30 crc kubenswrapper[4756]: I0203 09:28:30.990303 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" podStartSLOduration=4.990280886 podStartE2EDuration="4.990280886s" podCreationTimestamp="2026-02-03 09:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:30.983578243 +0000 UTC m=+1102.134045628" watchObservedRunningTime="2026-02-03 09:28:30.990280886 +0000 UTC m=+1102.140748261" Feb 03 09:28:31 crc kubenswrapper[4756]: I0203 09:28:31.461896 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:31 crc kubenswrapper[4756]: I0203 09:28:31.982823 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"19ccb5fd-4e2d-4cc8-b776-161687b99eaa","Type":"ContainerStarted","Data":"86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30"} Feb 03 09:28:31 crc kubenswrapper[4756]: I0203 09:28:31.987882 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16c4732e-295a-495d-a2af-e10b2c71a10f","Type":"ContainerStarted","Data":"0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef"} Feb 03 09:28:31 crc kubenswrapper[4756]: I0203 09:28:31.994133 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdf446f5c-8xnr7" event={"ID":"7fce406f-4aff-48f6-980a-64a65ede0c74","Type":"ContainerStarted","Data":"04ccca2b9393d6da33858a8f16a4af917eda96771ed9582ab23d1e05526c433b"} Feb 03 09:28:31 crc kubenswrapper[4756]: I0203 09:28:31.994176 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdf446f5c-8xnr7" event={"ID":"7fce406f-4aff-48f6-980a-64a65ede0c74","Type":"ContainerStarted","Data":"ab6dc282d7fe8bff4c4f1ba921f07b47a6680b5855d525ab5d7e63d20f3b24c6"} Feb 03 09:28:31 crc kubenswrapper[4756]: I0203 09:28:31.994884 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:28:32 crc kubenswrapper[4756]: I0203 09:28:31.998117 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerStarted","Data":"d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76"} Feb 03 09:28:32 crc kubenswrapper[4756]: I0203 09:28:32.008291 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.008272127 podStartE2EDuration="7.008272127s" podCreationTimestamp="2026-02-03 09:28:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:32.0032271 +0000 UTC m=+1103.153694475" watchObservedRunningTime="2026-02-03 09:28:32.008272127 +0000 UTC m=+1103.158739502" Feb 03 09:28:32 crc kubenswrapper[4756]: I0203 09:28:32.035491 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cdf446f5c-8xnr7" podStartSLOduration=3.035473498 podStartE2EDuration="3.035473498s" podCreationTimestamp="2026-02-03 09:28:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:32.025954891 +0000 UTC m=+1103.176422266" watchObservedRunningTime="2026-02-03 09:28:32.035473498 +0000 UTC m=+1103.185940873" Feb 03 09:28:33 crc kubenswrapper[4756]: I0203 09:28:33.006919 4756 generic.go:334] "Generic (PLEG): container finished" podID="79b123a2-b4b7-4233-98e5-538664495935" containerID="0439d59a66669486fb155370afc9837c626a409281f0397983d81cb77366332d" exitCode=0 Feb 03 09:28:33 crc kubenswrapper[4756]: I0203 09:28:33.007001 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7dmqr" event={"ID":"79b123a2-b4b7-4233-98e5-538664495935","Type":"ContainerDied","Data":"0439d59a66669486fb155370afc9837c626a409281f0397983d81cb77366332d"} Feb 03 09:28:33 crc kubenswrapper[4756]: I0203 09:28:33.010854 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16c4732e-295a-495d-a2af-e10b2c71a10f","Type":"ContainerStarted","Data":"5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b"} Feb 03 09:28:33 crc kubenswrapper[4756]: I0203 09:28:33.058532 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.058511516 podStartE2EDuration="7.058511516s" podCreationTimestamp="2026-02-03 09:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:33.051695723 +0000 UTC m=+1104.202163108" watchObservedRunningTime="2026-02-03 09:28:33.058511516 +0000 UTC m=+1104.208978891" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.042931 4756 generic.go:334] "Generic (PLEG): container finished" podID="45100e47-7ef9-499c-b831-ea2994af9403" containerID="66e2f1708ffff8fa5abe63648080d850074bd13d775bd047ec3c99fca2a25235" exitCode=0 Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.046926 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x9wrg" event={"ID":"45100e47-7ef9-499c-b831-ea2994af9403","Type":"ContainerDied","Data":"66e2f1708ffff8fa5abe63648080d850074bd13d775bd047ec3c99fca2a25235"} Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.509021 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7dmqr" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.690299 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6684\" (UniqueName: \"kubernetes.io/projected/79b123a2-b4b7-4233-98e5-538664495935-kube-api-access-b6684\") pod \"79b123a2-b4b7-4233-98e5-538664495935\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.690713 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-config-data\") pod \"79b123a2-b4b7-4233-98e5-538664495935\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.690743 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b123a2-b4b7-4233-98e5-538664495935-logs\") pod \"79b123a2-b4b7-4233-98e5-538664495935\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.690857 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-scripts\") pod \"79b123a2-b4b7-4233-98e5-538664495935\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.690909 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-combined-ca-bundle\") pod \"79b123a2-b4b7-4233-98e5-538664495935\" (UID: \"79b123a2-b4b7-4233-98e5-538664495935\") " Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.692232 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79b123a2-b4b7-4233-98e5-538664495935-logs" (OuterVolumeSpecName: "logs") pod "79b123a2-b4b7-4233-98e5-538664495935" (UID: "79b123a2-b4b7-4233-98e5-538664495935"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.700351 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-scripts" (OuterVolumeSpecName: "scripts") pod "79b123a2-b4b7-4233-98e5-538664495935" (UID: "79b123a2-b4b7-4233-98e5-538664495935"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.705952 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79b123a2-b4b7-4233-98e5-538664495935-kube-api-access-b6684" (OuterVolumeSpecName: "kube-api-access-b6684") pod "79b123a2-b4b7-4233-98e5-538664495935" (UID: "79b123a2-b4b7-4233-98e5-538664495935"). InnerVolumeSpecName "kube-api-access-b6684". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.742881 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-config-data" (OuterVolumeSpecName: "config-data") pod "79b123a2-b4b7-4233-98e5-538664495935" (UID: "79b123a2-b4b7-4233-98e5-538664495935"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.756787 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79b123a2-b4b7-4233-98e5-538664495935" (UID: "79b123a2-b4b7-4233-98e5-538664495935"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.794284 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.794559 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.794659 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6684\" (UniqueName: \"kubernetes.io/projected/79b123a2-b4b7-4233-98e5-538664495935-kube-api-access-b6684\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.794741 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79b123a2-b4b7-4233-98e5-538664495935-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:34 crc kubenswrapper[4756]: I0203 09:28:34.794824 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79b123a2-b4b7-4233-98e5-538664495935-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.054397 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7dmqr" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.054478 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7dmqr" event={"ID":"79b123a2-b4b7-4233-98e5-538664495935","Type":"ContainerDied","Data":"1e4c967f95e1f77088b1b0487275a07a80638bd6da30ee1d01e717ea271be80d"} Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.056274 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e4c967f95e1f77088b1b0487275a07a80638bd6da30ee1d01e717ea271be80d" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.174362 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ccf85886-5xql4"] Feb 03 09:28:35 crc kubenswrapper[4756]: E0203 09:28:35.175080 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b123a2-b4b7-4233-98e5-538664495935" containerName="placement-db-sync" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.175339 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b123a2-b4b7-4233-98e5-538664495935" containerName="placement-db-sync" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.175617 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="79b123a2-b4b7-4233-98e5-538664495935" containerName="placement-db-sync" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.179035 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.187065 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6ptgf" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.187428 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.187659 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.187914 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.188575 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.225928 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ccf85886-5xql4"] Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.306030 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-scripts\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.306126 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltcjc\" (UniqueName: \"kubernetes.io/projected/73d9e20e-cd45-437f-a2b5-a82370ed1038-kube-api-access-ltcjc\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.306190 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-internal-tls-certs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.306220 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-public-tls-certs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.307591 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-combined-ca-bundle\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.307640 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-config-data\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.307745 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73d9e20e-cd45-437f-a2b5-a82370ed1038-logs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.411660 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-combined-ca-bundle\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.411722 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-config-data\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.413329 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73d9e20e-cd45-437f-a2b5-a82370ed1038-logs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.411843 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73d9e20e-cd45-437f-a2b5-a82370ed1038-logs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.414779 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-scripts\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.414916 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltcjc\" (UniqueName: \"kubernetes.io/projected/73d9e20e-cd45-437f-a2b5-a82370ed1038-kube-api-access-ltcjc\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.415017 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-internal-tls-certs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.415060 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-public-tls-certs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.417687 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-combined-ca-bundle\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.427699 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-public-tls-certs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.432793 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-internal-tls-certs\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.433195 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-scripts\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.436236 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltcjc\" (UniqueName: \"kubernetes.io/projected/73d9e20e-cd45-437f-a2b5-a82370ed1038-kube-api-access-ltcjc\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.436508 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-config-data\") pod \"placement-ccf85886-5xql4\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:35 crc kubenswrapper[4756]: I0203 09:28:35.505854 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:36 crc kubenswrapper[4756]: I0203 09:28:36.311814 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 03 09:28:36 crc kubenswrapper[4756]: I0203 09:28:36.312188 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 03 09:28:36 crc kubenswrapper[4756]: I0203 09:28:36.425126 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 03 09:28:36 crc kubenswrapper[4756]: I0203 09:28:36.430398 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 03 09:28:36 crc kubenswrapper[4756]: I0203 09:28:36.491140 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:36 crc kubenswrapper[4756]: I0203 09:28:36.491197 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:36 crc kubenswrapper[4756]: I0203 09:28:36.527108 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:36 crc kubenswrapper[4756]: I0203 09:28:36.535412 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:37 crc kubenswrapper[4756]: I0203 09:28:37.080431 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 03 09:28:37 crc kubenswrapper[4756]: I0203 09:28:37.080734 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:37 crc kubenswrapper[4756]: I0203 09:28:37.080749 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:37 crc kubenswrapper[4756]: I0203 09:28:37.080758 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 03 09:28:37 crc kubenswrapper[4756]: I0203 09:28:37.175620 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:28:37 crc kubenswrapper[4756]: I0203 09:28:37.244570 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-vxpwc"] Feb 03 09:28:37 crc kubenswrapper[4756]: I0203 09:28:37.244799 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" podUID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerName="dnsmasq-dns" containerID="cri-o://b5f03b1082932810dc6dbce64c10649f12dca9ed05c19c3646acf5a50d9c7e15" gracePeriod=10 Feb 03 09:28:38 crc kubenswrapper[4756]: I0203 09:28:38.097982 4756 generic.go:334] "Generic (PLEG): container finished" podID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerID="b5f03b1082932810dc6dbce64c10649f12dca9ed05c19c3646acf5a50d9c7e15" exitCode=0 Feb 03 09:28:38 crc kubenswrapper[4756]: I0203 09:28:38.098054 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" event={"ID":"e1c29293-41e7-4617-a530-c0f0151d3e55","Type":"ContainerDied","Data":"b5f03b1082932810dc6dbce64c10649f12dca9ed05c19c3646acf5a50d9c7e15"} Feb 03 09:28:38 crc kubenswrapper[4756]: I0203 09:28:38.164707 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:38 crc kubenswrapper[4756]: I0203 09:28:38.164750 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:28:38 crc kubenswrapper[4756]: I0203 09:28:38.290641 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:38 crc kubenswrapper[4756]: I0203 09:28:38.290808 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.360791 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" podUID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.524713 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.698866 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-credential-keys\") pod \"45100e47-7ef9-499c-b831-ea2994af9403\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.698905 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-fernet-keys\") pod \"45100e47-7ef9-499c-b831-ea2994af9403\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.698984 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77kcm\" (UniqueName: \"kubernetes.io/projected/45100e47-7ef9-499c-b831-ea2994af9403-kube-api-access-77kcm\") pod \"45100e47-7ef9-499c-b831-ea2994af9403\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.699051 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-combined-ca-bundle\") pod \"45100e47-7ef9-499c-b831-ea2994af9403\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.699116 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-scripts\") pod \"45100e47-7ef9-499c-b831-ea2994af9403\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.699148 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-config-data\") pod \"45100e47-7ef9-499c-b831-ea2994af9403\" (UID: \"45100e47-7ef9-499c-b831-ea2994af9403\") " Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.748231 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "45100e47-7ef9-499c-b831-ea2994af9403" (UID: "45100e47-7ef9-499c-b831-ea2994af9403"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.766306 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-scripts" (OuterVolumeSpecName: "scripts") pod "45100e47-7ef9-499c-b831-ea2994af9403" (UID: "45100e47-7ef9-499c-b831-ea2994af9403"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.766930 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "45100e47-7ef9-499c-b831-ea2994af9403" (UID: "45100e47-7ef9-499c-b831-ea2994af9403"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.802943 4756 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.802983 4756 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.802994 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.803963 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45100e47-7ef9-499c-b831-ea2994af9403-kube-api-access-77kcm" (OuterVolumeSpecName: "kube-api-access-77kcm") pod "45100e47-7ef9-499c-b831-ea2994af9403" (UID: "45100e47-7ef9-499c-b831-ea2994af9403"). InnerVolumeSpecName "kube-api-access-77kcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.845920 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45100e47-7ef9-499c-b831-ea2994af9403" (UID: "45100e47-7ef9-499c-b831-ea2994af9403"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.850908 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-config-data" (OuterVolumeSpecName: "config-data") pod "45100e47-7ef9-499c-b831-ea2994af9403" (UID: "45100e47-7ef9-499c-b831-ea2994af9403"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.904843 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.904876 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45100e47-7ef9-499c-b831-ea2994af9403-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:39 crc kubenswrapper[4756]: I0203 09:28:39.904887 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77kcm\" (UniqueName: \"kubernetes.io/projected/45100e47-7ef9-499c-b831-ea2994af9403-kube-api-access-77kcm\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.010457 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.010807 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.018292 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.018380 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.056660 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.065954 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.140657 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x9wrg" event={"ID":"45100e47-7ef9-499c-b831-ea2994af9403","Type":"ContainerDied","Data":"8f1a01d83995e5795d989c6f88a137a96ae6a2487ac2e6e98d1562d1260552ac"} Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.140696 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f1a01d83995e5795d989c6f88a137a96ae6a2487ac2e6e98d1562d1260552ac" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.140707 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x9wrg" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.152762 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.153183 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-vxpwc" event={"ID":"e1c29293-41e7-4617-a530-c0f0151d3e55","Type":"ContainerDied","Data":"feccabcc1db2e5486baac385e84fc91db24f838b5b58fd67c525233986700f56"} Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.153215 4756 scope.go:117] "RemoveContainer" containerID="b5f03b1082932810dc6dbce64c10649f12dca9ed05c19c3646acf5a50d9c7e15" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.180580 4756 scope.go:117] "RemoveContainer" containerID="3ef7348df72bc7a80d4d77cceaefba5c8a8a103f36fb03276e5b64034b65b01a" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.197758 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.224553 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72rkb\" (UniqueName: \"kubernetes.io/projected/e1c29293-41e7-4617-a530-c0f0151d3e55-kube-api-access-72rkb\") pod \"e1c29293-41e7-4617-a530-c0f0151d3e55\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.224649 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-sb\") pod \"e1c29293-41e7-4617-a530-c0f0151d3e55\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.224699 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-swift-storage-0\") pod \"e1c29293-41e7-4617-a530-c0f0151d3e55\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.224776 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-config\") pod \"e1c29293-41e7-4617-a530-c0f0151d3e55\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.224836 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-nb\") pod \"e1c29293-41e7-4617-a530-c0f0151d3e55\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.224859 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-svc\") pod \"e1c29293-41e7-4617-a530-c0f0151d3e55\" (UID: \"e1c29293-41e7-4617-a530-c0f0151d3e55\") " Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.261618 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c29293-41e7-4617-a530-c0f0151d3e55-kube-api-access-72rkb" (OuterVolumeSpecName: "kube-api-access-72rkb") pod "e1c29293-41e7-4617-a530-c0f0151d3e55" (UID: "e1c29293-41e7-4617-a530-c0f0151d3e55"). InnerVolumeSpecName "kube-api-access-72rkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.315108 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ccf85886-5xql4"] Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.332982 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72rkb\" (UniqueName: \"kubernetes.io/projected/e1c29293-41e7-4617-a530-c0f0151d3e55-kube-api-access-72rkb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:40 crc kubenswrapper[4756]: W0203 09:28:40.347561 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73d9e20e_cd45_437f_a2b5_a82370ed1038.slice/crio-6ae9623200e35a3e332b68086aaa16453aea8843aa4f57fb16eb7a86f24e33a9 WatchSource:0}: Error finding container 6ae9623200e35a3e332b68086aaa16453aea8843aa4f57fb16eb7a86f24e33a9: Status 404 returned error can't find the container with id 6ae9623200e35a3e332b68086aaa16453aea8843aa4f57fb16eb7a86f24e33a9 Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.373538 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e1c29293-41e7-4617-a530-c0f0151d3e55" (UID: "e1c29293-41e7-4617-a530-c0f0151d3e55"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.385068 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e1c29293-41e7-4617-a530-c0f0151d3e55" (UID: "e1c29293-41e7-4617-a530-c0f0151d3e55"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.426175 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e1c29293-41e7-4617-a530-c0f0151d3e55" (UID: "e1c29293-41e7-4617-a530-c0f0151d3e55"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.434477 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.434509 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.434521 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.478284 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e1c29293-41e7-4617-a530-c0f0151d3e55" (UID: "e1c29293-41e7-4617-a530-c0f0151d3e55"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.478715 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-config" (OuterVolumeSpecName: "config") pod "e1c29293-41e7-4617-a530-c0f0151d3e55" (UID: "e1c29293-41e7-4617-a530-c0f0151d3e55"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.538992 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.539028 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1c29293-41e7-4617-a530-c0f0151d3e55-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.687431 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-765cf46959-4rcst"] Feb 03 09:28:40 crc kubenswrapper[4756]: E0203 09:28:40.689653 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45100e47-7ef9-499c-b831-ea2994af9403" containerName="keystone-bootstrap" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.689676 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="45100e47-7ef9-499c-b831-ea2994af9403" containerName="keystone-bootstrap" Feb 03 09:28:40 crc kubenswrapper[4756]: E0203 09:28:40.689723 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerName="init" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.689733 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerName="init" Feb 03 09:28:40 crc kubenswrapper[4756]: E0203 09:28:40.689757 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerName="dnsmasq-dns" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.689767 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerName="dnsmasq-dns" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.690156 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="45100e47-7ef9-499c-b831-ea2994af9403" containerName="keystone-bootstrap" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.690192 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c29293-41e7-4617-a530-c0f0151d3e55" containerName="dnsmasq-dns" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.690890 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-765cf46959-4rcst"] Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.690986 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.695775 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.695846 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.696118 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.696337 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.696557 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fh27s" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.696797 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.745295 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-internal-tls-certs\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.745456 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zbhc\" (UniqueName: \"kubernetes.io/projected/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-kube-api-access-6zbhc\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.745523 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-config-data\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.745547 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-public-tls-certs\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.745577 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-combined-ca-bundle\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.745612 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-fernet-keys\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.745657 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-scripts\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.745770 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-credential-keys\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.847811 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-credential-keys\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.847932 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-internal-tls-certs\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.847998 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zbhc\" (UniqueName: \"kubernetes.io/projected/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-kube-api-access-6zbhc\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.848021 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-config-data\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.848043 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-public-tls-certs\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.848071 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-combined-ca-bundle\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.848095 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-fernet-keys\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.848143 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-scripts\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.853347 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-scripts\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.862283 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-config-data\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.864158 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-combined-ca-bundle\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.864844 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-fernet-keys\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.869118 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-public-tls-certs\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.872856 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-credential-keys\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.874050 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-internal-tls-certs\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.876140 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zbhc\" (UniqueName: \"kubernetes.io/projected/37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf-kube-api-access-6zbhc\") pod \"keystone-765cf46959-4rcst\" (UID: \"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf\") " pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.904230 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.938504 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-vxpwc"] Feb 03 09:28:40 crc kubenswrapper[4756]: I0203 09:28:40.948261 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-vxpwc"] Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.169235 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ccf85886-5xql4" event={"ID":"73d9e20e-cd45-437f-a2b5-a82370ed1038","Type":"ContainerStarted","Data":"e19a82576e77552afbabff92a0e0ebf8849f51dc4b7c82ea480989825c3b0709"} Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.169632 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ccf85886-5xql4" event={"ID":"73d9e20e-cd45-437f-a2b5-a82370ed1038","Type":"ContainerStarted","Data":"c8a4e13ff5983b0dc98fd0638ead6bbbfcd0258604e9de29739a075045be5e83"} Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.169650 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ccf85886-5xql4" event={"ID":"73d9e20e-cd45-437f-a2b5-a82370ed1038","Type":"ContainerStarted","Data":"6ae9623200e35a3e332b68086aaa16453aea8843aa4f57fb16eb7a86f24e33a9"} Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.170887 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.170915 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-ccf85886-5xql4" Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.178804 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerStarted","Data":"fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c"} Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.179914 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cbcxp" event={"ID":"1d6582a5-c7dc-4444-924e-9b977a3d4378","Type":"ContainerStarted","Data":"5fe5f459c0dd48a21d9f0786435c92fa5dad2a769a3b44126c85575ac0fd1558"} Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.205621 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-ccf85886-5xql4" podStartSLOduration=6.205602677 podStartE2EDuration="6.205602677s" podCreationTimestamp="2026-02-03 09:28:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:41.196646856 +0000 UTC m=+1112.347114251" watchObservedRunningTime="2026-02-03 09:28:41.205602677 +0000 UTC m=+1112.356070052" Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.222650 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-cbcxp" podStartSLOduration=4.059247078 podStartE2EDuration="43.222625739s" podCreationTimestamp="2026-02-03 09:27:58 +0000 UTC" firstStartedPulling="2026-02-03 09:28:01.018230948 +0000 UTC m=+1072.168698323" lastFinishedPulling="2026-02-03 09:28:40.181609609 +0000 UTC m=+1111.332076984" observedRunningTime="2026-02-03 09:28:41.217780847 +0000 UTC m=+1112.368248232" watchObservedRunningTime="2026-02-03 09:28:41.222625739 +0000 UTC m=+1112.373093114" Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.459169 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-765cf46959-4rcst"] Feb 03 09:28:41 crc kubenswrapper[4756]: I0203 09:28:41.639576 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1c29293-41e7-4617-a530-c0f0151d3e55" path="/var/lib/kubelet/pods/e1c29293-41e7-4617-a530-c0f0151d3e55/volumes" Feb 03 09:28:42 crc kubenswrapper[4756]: I0203 09:28:42.188795 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4ld7v" event={"ID":"7dee8740-0310-450c-88fd-363ce0a26d4b","Type":"ContainerStarted","Data":"77986f06d0b2cfa7481468c9ca08776ae2dcce0ceb8ccdc36b9276a581777f7c"} Feb 03 09:28:42 crc kubenswrapper[4756]: I0203 09:28:42.192754 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-765cf46959-4rcst" event={"ID":"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf","Type":"ContainerStarted","Data":"d42b3e7912cb824333ec8e801f826987b17fcf3260802946934e7d3e32dee920"} Feb 03 09:28:42 crc kubenswrapper[4756]: I0203 09:28:42.192790 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:28:42 crc kubenswrapper[4756]: I0203 09:28:42.192800 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-765cf46959-4rcst" event={"ID":"37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf","Type":"ContainerStarted","Data":"8c9f7d32e508d34dae8fd7d16fcc9392ac6023d391771687afe1e34666f6dbb5"} Feb 03 09:28:42 crc kubenswrapper[4756]: I0203 09:28:42.210631 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-4ld7v" podStartSLOduration=3.582258602 podStartE2EDuration="44.210611581s" podCreationTimestamp="2026-02-03 09:27:58 +0000 UTC" firstStartedPulling="2026-02-03 09:28:00.441258624 +0000 UTC m=+1071.591725999" lastFinishedPulling="2026-02-03 09:28:41.069611603 +0000 UTC m=+1112.220078978" observedRunningTime="2026-02-03 09:28:42.207977818 +0000 UTC m=+1113.358445193" watchObservedRunningTime="2026-02-03 09:28:42.210611581 +0000 UTC m=+1113.361078956" Feb 03 09:28:42 crc kubenswrapper[4756]: I0203 09:28:42.234165 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-765cf46959-4rcst" podStartSLOduration=2.234142956 podStartE2EDuration="2.234142956s" podCreationTimestamp="2026-02-03 09:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:42.232070722 +0000 UTC m=+1113.382538097" watchObservedRunningTime="2026-02-03 09:28:42.234142956 +0000 UTC m=+1113.384610331" Feb 03 09:28:43 crc kubenswrapper[4756]: I0203 09:28:43.566677 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:28:43 crc kubenswrapper[4756]: I0203 09:28:43.566758 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:28:45 crc kubenswrapper[4756]: I0203 09:28:45.234899 4756 generic.go:334] "Generic (PLEG): container finished" podID="1d6582a5-c7dc-4444-924e-9b977a3d4378" containerID="5fe5f459c0dd48a21d9f0786435c92fa5dad2a769a3b44126c85575ac0fd1558" exitCode=0 Feb 03 09:28:45 crc kubenswrapper[4756]: I0203 09:28:45.234985 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cbcxp" event={"ID":"1d6582a5-c7dc-4444-924e-9b977a3d4378","Type":"ContainerDied","Data":"5fe5f459c0dd48a21d9f0786435c92fa5dad2a769a3b44126c85575ac0fd1558"} Feb 03 09:28:47 crc kubenswrapper[4756]: I0203 09:28:47.260388 4756 generic.go:334] "Generic (PLEG): container finished" podID="7dee8740-0310-450c-88fd-363ce0a26d4b" containerID="77986f06d0b2cfa7481468c9ca08776ae2dcce0ceb8ccdc36b9276a581777f7c" exitCode=0 Feb 03 09:28:47 crc kubenswrapper[4756]: I0203 09:28:47.260467 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4ld7v" event={"ID":"7dee8740-0310-450c-88fd-363ce0a26d4b","Type":"ContainerDied","Data":"77986f06d0b2cfa7481468c9ca08776ae2dcce0ceb8ccdc36b9276a581777f7c"} Feb 03 09:28:48 crc kubenswrapper[4756]: I0203 09:28:48.167319 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-546db499dd-c5f8j" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Feb 03 09:28:48 crc kubenswrapper[4756]: I0203 09:28:48.292563 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6d95f7b7bb-r6fd8" podUID="c34fbdee-8998-4e2d-bc98-c5c73ff35160" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.288156 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4ld7v" event={"ID":"7dee8740-0310-450c-88fd-363ce0a26d4b","Type":"ContainerDied","Data":"0257aa128c7f4b4ec0062d1916da66e456f9e3cd26a33a7a371317d49b3bc992"} Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.288409 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0257aa128c7f4b4ec0062d1916da66e456f9e3cd26a33a7a371317d49b3bc992" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.290368 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-cbcxp" event={"ID":"1d6582a5-c7dc-4444-924e-9b977a3d4378","Type":"ContainerDied","Data":"d194f31f8ea72988ec864a0b3fe3aa2754b8ae85a79e9f08732d678a0bee3f9f"} Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.290392 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d194f31f8ea72988ec864a0b3fe3aa2754b8ae85a79e9f08732d678a0bee3f9f" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.333032 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.338986 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445328 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8gzp\" (UniqueName: \"kubernetes.io/projected/1d6582a5-c7dc-4444-924e-9b977a3d4378-kube-api-access-z8gzp\") pod \"1d6582a5-c7dc-4444-924e-9b977a3d4378\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445413 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-config-data\") pod \"7dee8740-0310-450c-88fd-363ce0a26d4b\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445498 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-combined-ca-bundle\") pod \"1d6582a5-c7dc-4444-924e-9b977a3d4378\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445525 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7dee8740-0310-450c-88fd-363ce0a26d4b-etc-machine-id\") pod \"7dee8740-0310-450c-88fd-363ce0a26d4b\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445549 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-combined-ca-bundle\") pod \"7dee8740-0310-450c-88fd-363ce0a26d4b\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445588 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq64w\" (UniqueName: \"kubernetes.io/projected/7dee8740-0310-450c-88fd-363ce0a26d4b-kube-api-access-hq64w\") pod \"7dee8740-0310-450c-88fd-363ce0a26d4b\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445603 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-db-sync-config-data\") pod \"7dee8740-0310-450c-88fd-363ce0a26d4b\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445632 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-db-sync-config-data\") pod \"1d6582a5-c7dc-4444-924e-9b977a3d4378\" (UID: \"1d6582a5-c7dc-4444-924e-9b977a3d4378\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.445706 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-scripts\") pod \"7dee8740-0310-450c-88fd-363ce0a26d4b\" (UID: \"7dee8740-0310-450c-88fd-363ce0a26d4b\") " Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.448092 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7dee8740-0310-450c-88fd-363ce0a26d4b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7dee8740-0310-450c-88fd-363ce0a26d4b" (UID: "7dee8740-0310-450c-88fd-363ce0a26d4b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.452235 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d6582a5-c7dc-4444-924e-9b977a3d4378-kube-api-access-z8gzp" (OuterVolumeSpecName: "kube-api-access-z8gzp") pod "1d6582a5-c7dc-4444-924e-9b977a3d4378" (UID: "1d6582a5-c7dc-4444-924e-9b977a3d4378"). InnerVolumeSpecName "kube-api-access-z8gzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.452325 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dee8740-0310-450c-88fd-363ce0a26d4b-kube-api-access-hq64w" (OuterVolumeSpecName: "kube-api-access-hq64w") pod "7dee8740-0310-450c-88fd-363ce0a26d4b" (UID: "7dee8740-0310-450c-88fd-363ce0a26d4b"). InnerVolumeSpecName "kube-api-access-hq64w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.453479 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7dee8740-0310-450c-88fd-363ce0a26d4b" (UID: "7dee8740-0310-450c-88fd-363ce0a26d4b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.453782 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1d6582a5-c7dc-4444-924e-9b977a3d4378" (UID: "1d6582a5-c7dc-4444-924e-9b977a3d4378"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.454025 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-scripts" (OuterVolumeSpecName: "scripts") pod "7dee8740-0310-450c-88fd-363ce0a26d4b" (UID: "7dee8740-0310-450c-88fd-363ce0a26d4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.486744 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dee8740-0310-450c-88fd-363ce0a26d4b" (UID: "7dee8740-0310-450c-88fd-363ce0a26d4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.489248 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d6582a5-c7dc-4444-924e-9b977a3d4378" (UID: "1d6582a5-c7dc-4444-924e-9b977a3d4378"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.517691 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-config-data" (OuterVolumeSpecName: "config-data") pod "7dee8740-0310-450c-88fd-363ce0a26d4b" (UID: "7dee8740-0310-450c-88fd-363ce0a26d4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548021 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8gzp\" (UniqueName: \"kubernetes.io/projected/1d6582a5-c7dc-4444-924e-9b977a3d4378-kube-api-access-z8gzp\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548064 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548073 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548103 4756 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7dee8740-0310-450c-88fd-363ce0a26d4b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548114 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548130 4756 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548146 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq64w\" (UniqueName: \"kubernetes.io/projected/7dee8740-0310-450c-88fd-363ce0a26d4b-kube-api-access-hq64w\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548184 4756 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d6582a5-c7dc-4444-924e-9b977a3d4378-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:50 crc kubenswrapper[4756]: I0203 09:28:50.548193 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dee8740-0310-450c-88fd-363ce0a26d4b-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.302295 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerStarted","Data":"623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95"} Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.302308 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-cbcxp" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.302428 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4ld7v" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.302861 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="ceilometer-central-agent" containerID="cri-o://72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273" gracePeriod=30 Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.303151 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="proxy-httpd" containerID="cri-o://623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95" gracePeriod=30 Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.303253 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="sg-core" containerID="cri-o://fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c" gracePeriod=30 Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.303311 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="ceilometer-notification-agent" containerID="cri-o://d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76" gracePeriod=30 Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.323866 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.791974111 podStartE2EDuration="53.323851891s" podCreationTimestamp="2026-02-03 09:27:58 +0000 UTC" firstStartedPulling="2026-02-03 09:28:00.909831909 +0000 UTC m=+1072.060299284" lastFinishedPulling="2026-02-03 09:28:50.441709689 +0000 UTC m=+1121.592177064" observedRunningTime="2026-02-03 09:28:51.321733064 +0000 UTC m=+1122.472200439" watchObservedRunningTime="2026-02-03 09:28:51.323851891 +0000 UTC m=+1122.474319266" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.666514 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:28:51 crc kubenswrapper[4756]: E0203 09:28:51.667273 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6582a5-c7dc-4444-924e-9b977a3d4378" containerName="barbican-db-sync" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.667292 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6582a5-c7dc-4444-924e-9b977a3d4378" containerName="barbican-db-sync" Feb 03 09:28:51 crc kubenswrapper[4756]: E0203 09:28:51.667329 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dee8740-0310-450c-88fd-363ce0a26d4b" containerName="cinder-db-sync" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.667338 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dee8740-0310-450c-88fd-363ce0a26d4b" containerName="cinder-db-sync" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.667605 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d6582a5-c7dc-4444-924e-9b977a3d4378" containerName="barbican-db-sync" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.667619 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dee8740-0310-450c-88fd-363ce0a26d4b" containerName="cinder-db-sync" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.668865 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.674984 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wc6nw" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.675078 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.695352 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.695539 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.695541 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-68b68864ff-4znh2"] Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.697336 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.702970 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.703202 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.703206 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8hfm6" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.708980 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5948f79bd-cl5cq"] Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.770659 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.773902 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-config-data-custom\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.773948 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3dd59369-6dbb-4877-886c-76f19f87b90a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.773978 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fssjn\" (UniqueName: \"kubernetes.io/projected/8b049610-c8e6-40bd-98cb-138f42f1c35f-kube-api-access-fssjn\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.774018 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.774066 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj44g\" (UniqueName: \"kubernetes.io/projected/3dd59369-6dbb-4877-886c-76f19f87b90a-kube-api-access-jj44g\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.774092 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b049610-c8e6-40bd-98cb-138f42f1c35f-logs\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.774116 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-config-data\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.774135 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-scripts\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.774151 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-combined-ca-bundle\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.774173 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.774196 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.782149 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.835622 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876115 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9cck\" (UniqueName: \"kubernetes.io/projected/c34e66d1-0527-4f40-a676-1ac69c4d723f-kube-api-access-j9cck\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876244 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-config-data-custom\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876277 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3dd59369-6dbb-4877-886c-76f19f87b90a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876300 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-combined-ca-bundle\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876318 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fssjn\" (UniqueName: \"kubernetes.io/projected/8b049610-c8e6-40bd-98cb-138f42f1c35f-kube-api-access-fssjn\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876352 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876366 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c34e66d1-0527-4f40-a676-1ac69c4d723f-logs\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876400 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-config-data\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876420 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj44g\" (UniqueName: \"kubernetes.io/projected/3dd59369-6dbb-4877-886c-76f19f87b90a-kube-api-access-jj44g\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876453 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b049610-c8e6-40bd-98cb-138f42f1c35f-logs\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876472 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-config-data\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876486 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-scripts\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876501 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-config-data-custom\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876515 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-combined-ca-bundle\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876530 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.876549 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.886468 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-combined-ca-bundle\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.886680 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3dd59369-6dbb-4877-886c-76f19f87b90a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.886745 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b049610-c8e6-40bd-98cb-138f42f1c35f-logs\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.887852 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-config-data-custom\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.888064 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-scripts\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.898346 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68b68864ff-4znh2"] Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.899244 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.905387 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b049610-c8e6-40bd-98cb-138f42f1c35f-config-data\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.911975 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj44g\" (UniqueName: \"kubernetes.io/projected/3dd59369-6dbb-4877-886c-76f19f87b90a-kube-api-access-jj44g\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.912571 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.915146 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.916090 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fssjn\" (UniqueName: \"kubernetes.io/projected/8b049610-c8e6-40bd-98cb-138f42f1c35f-kube-api-access-fssjn\") pod \"barbican-worker-68b68864ff-4znh2\" (UID: \"8b049610-c8e6-40bd-98cb-138f42f1c35f\") " pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.920553 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5948f79bd-cl5cq"] Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.978365 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-config-data-custom\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.978670 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9cck\" (UniqueName: \"kubernetes.io/projected/c34e66d1-0527-4f40-a676-1ac69c4d723f-kube-api-access-j9cck\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.978775 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-combined-ca-bundle\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.978859 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c34e66d1-0527-4f40-a676-1ac69c4d723f-logs\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.978943 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-config-data\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.984286 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c34e66d1-0527-4f40-a676-1ac69c4d723f-logs\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.985519 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-config-data-custom\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.997328 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 03 09:28:51 crc kubenswrapper[4756]: I0203 09:28:51.997637 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-config-data\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.002727 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34e66d1-0527-4f40-a676-1ac69c4d723f-combined-ca-bundle\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.036223 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9cck\" (UniqueName: \"kubernetes.io/projected/c34e66d1-0527-4f40-a676-1ac69c4d723f-kube-api-access-j9cck\") pod \"barbican-keystone-listener-5948f79bd-cl5cq\" (UID: \"c34e66d1-0527-4f40-a676-1ac69c4d723f\") " pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.046469 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-86dm8"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.047943 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.079948 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.080002 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6v2d\" (UniqueName: \"kubernetes.io/projected/deb43135-f681-4b70-99f7-28262816638e-kube-api-access-v6v2d\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.080035 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.080057 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-config\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.080102 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.080126 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.090735 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-86dm8"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.098164 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68b68864ff-4znh2" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.142798 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.159782 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-86dm8"] Feb 03 09:28:52 crc kubenswrapper[4756]: E0203 09:28:52.161283 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-v6v2d ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" podUID="deb43135-f681-4b70-99f7-28262816638e" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.172138 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-77mfv"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.174536 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.181822 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.181987 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.182077 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6v2d\" (UniqueName: \"kubernetes.io/projected/deb43135-f681-4b70-99f7-28262816638e-kube-api-access-v6v2d\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.182139 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.182170 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-config\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.182433 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.183604 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.184669 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.184669 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.185585 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-config\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.193685 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.202094 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-77mfv"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.217811 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6v2d\" (UniqueName: \"kubernetes.io/projected/deb43135-f681-4b70-99f7-28262816638e-kube-api-access-v6v2d\") pod \"dnsmasq-dns-848cf88cfc-86dm8\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.229326 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7b7c55c5fb-l7j2x"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.231172 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.233061 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.246143 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.251346 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.259571 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.259917 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b7c55c5fb-l7j2x"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.276334 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.283655 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-logs\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.283929 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.286951 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.287229 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j9xp\" (UniqueName: \"kubernetes.io/projected/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-kube-api-access-9j9xp\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.287340 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.287558 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data-custom\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.295748 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data-custom\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.296228 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.296371 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-svc\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.296577 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpj9w\" (UniqueName: \"kubernetes.io/projected/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-kube-api-access-tpj9w\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.296754 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-combined-ca-bundle\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.296882 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.296981 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.298195 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-config\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.298320 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6h8k\" (UniqueName: \"kubernetes.io/projected/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-kube-api-access-q6h8k\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.298424 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.298616 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-logs\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.298693 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-scripts\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.349607 4756 generic.go:334] "Generic (PLEG): container finished" podID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerID="623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95" exitCode=0 Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.349651 4756 generic.go:334] "Generic (PLEG): container finished" podID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerID="fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c" exitCode=2 Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.349661 4756 generic.go:334] "Generic (PLEG): container finished" podID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerID="72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273" exitCode=0 Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.349725 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.350408 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerDied","Data":"623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95"} Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.350755 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerDied","Data":"fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c"} Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.350769 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerDied","Data":"72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273"} Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.376655 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.408897 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-nb\") pod \"deb43135-f681-4b70-99f7-28262816638e\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.408949 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-swift-storage-0\") pod \"deb43135-f681-4b70-99f7-28262816638e\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409002 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-config\") pod \"deb43135-f681-4b70-99f7-28262816638e\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409087 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-sb\") pod \"deb43135-f681-4b70-99f7-28262816638e\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409176 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6v2d\" (UniqueName: \"kubernetes.io/projected/deb43135-f681-4b70-99f7-28262816638e-kube-api-access-v6v2d\") pod \"deb43135-f681-4b70-99f7-28262816638e\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409264 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-svc\") pod \"deb43135-f681-4b70-99f7-28262816638e\" (UID: \"deb43135-f681-4b70-99f7-28262816638e\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409607 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-svc\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409644 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpj9w\" (UniqueName: \"kubernetes.io/projected/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-kube-api-access-tpj9w\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409685 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-combined-ca-bundle\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409719 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409743 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409764 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6h8k\" (UniqueName: \"kubernetes.io/projected/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-kube-api-access-q6h8k\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409779 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-config\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409797 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409820 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-logs\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409843 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-scripts\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409868 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-logs\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409886 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409906 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409930 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j9xp\" (UniqueName: \"kubernetes.io/projected/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-kube-api-access-9j9xp\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409948 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409988 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data-custom\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.410021 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data-custom\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.410037 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.414609 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.415688 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-config\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.416252 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.416517 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-logs\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.421382 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.423317 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.409728 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "deb43135-f681-4b70-99f7-28262816638e" (UID: "deb43135-f681-4b70-99f7-28262816638e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.410094 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "deb43135-f681-4b70-99f7-28262816638e" (UID: "deb43135-f681-4b70-99f7-28262816638e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.410831 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-config" (OuterVolumeSpecName: "config") pod "deb43135-f681-4b70-99f7-28262816638e" (UID: "deb43135-f681-4b70-99f7-28262816638e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.411220 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "deb43135-f681-4b70-99f7-28262816638e" (UID: "deb43135-f681-4b70-99f7-28262816638e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.421327 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "deb43135-f681-4b70-99f7-28262816638e" (UID: "deb43135-f681-4b70-99f7-28262816638e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.424353 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-svc\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.425038 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-combined-ca-bundle\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.425298 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.425830 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deb43135-f681-4b70-99f7-28262816638e-kube-api-access-v6v2d" (OuterVolumeSpecName: "kube-api-access-v6v2d") pod "deb43135-f681-4b70-99f7-28262816638e" (UID: "deb43135-f681-4b70-99f7-28262816638e"). InnerVolumeSpecName "kube-api-access-v6v2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.426140 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-logs\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.426717 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-scripts\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.429961 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data-custom\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.430611 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.439270 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data-custom\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.446556 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.448745 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6h8k\" (UniqueName: \"kubernetes.io/projected/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-kube-api-access-q6h8k\") pod \"dnsmasq-dns-6578955fd5-77mfv\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.483625 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpj9w\" (UniqueName: \"kubernetes.io/projected/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-kube-api-access-tpj9w\") pod \"cinder-api-0\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.494659 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j9xp\" (UniqueName: \"kubernetes.io/projected/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-kube-api-access-9j9xp\") pod \"barbican-api-7b7c55c5fb-l7j2x\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.496883 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.512806 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.512844 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6v2d\" (UniqueName: \"kubernetes.io/projected/deb43135-f681-4b70-99f7-28262816638e-kube-api-access-v6v2d\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.512890 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.517768 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.517812 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.517829 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb43135-f681-4b70-99f7-28262816638e-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.588931 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.615300 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.637851 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.834627 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68b68864ff-4znh2"] Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.891164 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.934710 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-config-data\") pod \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.934834 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvlz9\" (UniqueName: \"kubernetes.io/projected/d7957443-ab24-47bf-a76e-0f0b65d9ceca-kube-api-access-kvlz9\") pod \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.934885 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-sg-core-conf-yaml\") pod \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.934925 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-run-httpd\") pod \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.934977 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-scripts\") pod \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.935054 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-log-httpd\") pod \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.935181 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-combined-ca-bundle\") pod \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\" (UID: \"d7957443-ab24-47bf-a76e-0f0b65d9ceca\") " Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.936666 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d7957443-ab24-47bf-a76e-0f0b65d9ceca" (UID: "d7957443-ab24-47bf-a76e-0f0b65d9ceca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.936729 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d7957443-ab24-47bf-a76e-0f0b65d9ceca" (UID: "d7957443-ab24-47bf-a76e-0f0b65d9ceca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.942218 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7957443-ab24-47bf-a76e-0f0b65d9ceca-kube-api-access-kvlz9" (OuterVolumeSpecName: "kube-api-access-kvlz9") pod "d7957443-ab24-47bf-a76e-0f0b65d9ceca" (UID: "d7957443-ab24-47bf-a76e-0f0b65d9ceca"). InnerVolumeSpecName "kube-api-access-kvlz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.943668 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-scripts" (OuterVolumeSpecName: "scripts") pod "d7957443-ab24-47bf-a76e-0f0b65d9ceca" (UID: "d7957443-ab24-47bf-a76e-0f0b65d9ceca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:52 crc kubenswrapper[4756]: I0203 09:28:52.973667 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d7957443-ab24-47bf-a76e-0f0b65d9ceca" (UID: "d7957443-ab24-47bf-a76e-0f0b65d9ceca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.020568 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5948f79bd-cl5cq"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.037602 4756 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.037684 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.037696 4756 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7957443-ab24-47bf-a76e-0f0b65d9ceca-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.037707 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvlz9\" (UniqueName: \"kubernetes.io/projected/d7957443-ab24-47bf-a76e-0f0b65d9ceca-kube-api-access-kvlz9\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.037721 4756 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.039527 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7957443-ab24-47bf-a76e-0f0b65d9ceca" (UID: "d7957443-ab24-47bf-a76e-0f0b65d9ceca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.079666 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-config-data" (OuterVolumeSpecName: "config-data") pod "d7957443-ab24-47bf-a76e-0f0b65d9ceca" (UID: "d7957443-ab24-47bf-a76e-0f0b65d9ceca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.139051 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.139102 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7957443-ab24-47bf-a76e-0f0b65d9ceca-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.236227 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-77mfv"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.353058 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.361636 4756 generic.go:334] "Generic (PLEG): container finished" podID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerID="d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76" exitCode=0 Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.361713 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerDied","Data":"d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76"} Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.361742 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7957443-ab24-47bf-a76e-0f0b65d9ceca","Type":"ContainerDied","Data":"d4b89e209e17e69db44a1e044eb3c394ecf8a4a217c0c6b6c1bd4d54716d1838"} Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.361760 4756 scope.go:117] "RemoveContainer" containerID="623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.361894 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.374716 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3dd59369-6dbb-4877-886c-76f19f87b90a","Type":"ContainerStarted","Data":"bbd3136bd4711b71de4852098f921d420d4f67898c0a25dff7afb002e3003bb6"} Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.376496 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" event={"ID":"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae","Type":"ContainerStarted","Data":"10d08e0a4eb1511a8f3d6db049319c60f27f248557a53da56e721af0b930e941"} Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.377869 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b68864ff-4znh2" event={"ID":"8b049610-c8e6-40bd-98cb-138f42f1c35f","Type":"ContainerStarted","Data":"2e3721043220a90a83dfb8b23555d398f5b06fdb564faea4b92128e65a7b9cde"} Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.383514 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-86dm8" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.384926 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" event={"ID":"c34e66d1-0527-4f40-a676-1ac69c4d723f","Type":"ContainerStarted","Data":"7aafe97ef81d1c606003a9cea6d9609cebc8be313e3ecc9754dd2d4e76f18c6d"} Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.387209 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b7c55c5fb-l7j2x"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.432554 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.468392 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.511320 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:28:53 crc kubenswrapper[4756]: E0203 09:28:53.511920 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="sg-core" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.511939 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="sg-core" Feb 03 09:28:53 crc kubenswrapper[4756]: E0203 09:28:53.511958 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="proxy-httpd" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.511966 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="proxy-httpd" Feb 03 09:28:53 crc kubenswrapper[4756]: E0203 09:28:53.511987 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="ceilometer-notification-agent" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.511994 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="ceilometer-notification-agent" Feb 03 09:28:53 crc kubenswrapper[4756]: E0203 09:28:53.512018 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="ceilometer-central-agent" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.512025 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="ceilometer-central-agent" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.512273 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="proxy-httpd" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.512299 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="ceilometer-central-agent" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.512322 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="ceilometer-notification-agent" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.512333 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" containerName="sg-core" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.516408 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.518585 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-86dm8"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.521069 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.521498 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.529658 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-86dm8"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.543693 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.553291 4756 scope.go:117] "RemoveContainer" containerID="fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.561814 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-run-httpd\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.563836 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.563923 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-log-httpd\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.565796 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-config-data\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.565849 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-scripts\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.566247 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrcd8\" (UniqueName: \"kubernetes.io/projected/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-kube-api-access-mrcd8\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.566312 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.627650 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7957443-ab24-47bf-a76e-0f0b65d9ceca" path="/var/lib/kubelet/pods/d7957443-ab24-47bf-a76e-0f0b65d9ceca/volumes" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.628677 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deb43135-f681-4b70-99f7-28262816638e" path="/var/lib/kubelet/pods/deb43135-f681-4b70-99f7-28262816638e/volumes" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.636905 4756 scope.go:117] "RemoveContainer" containerID="d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.673661 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-config-data\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.675322 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-scripts\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.675482 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrcd8\" (UniqueName: \"kubernetes.io/projected/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-kube-api-access-mrcd8\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.675529 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.675573 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-run-httpd\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.675646 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.675684 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-log-httpd\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.676300 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-run-httpd\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.676489 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-log-httpd\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.679550 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-config-data\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.682994 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-scripts\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.684519 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.687143 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.708933 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrcd8\" (UniqueName: \"kubernetes.io/projected/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-kube-api-access-mrcd8\") pod \"ceilometer-0\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.709271 4756 scope.go:117] "RemoveContainer" containerID="72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.817089 4756 scope.go:117] "RemoveContainer" containerID="623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95" Feb 03 09:28:53 crc kubenswrapper[4756]: E0203 09:28:53.820540 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95\": container with ID starting with 623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95 not found: ID does not exist" containerID="623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.820586 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95"} err="failed to get container status \"623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95\": rpc error: code = NotFound desc = could not find container \"623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95\": container with ID starting with 623fd94744506e44375bfab89d610c1ff8ee49f4769a1c667184faa2299a7d95 not found: ID does not exist" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.820616 4756 scope.go:117] "RemoveContainer" containerID="fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c" Feb 03 09:28:53 crc kubenswrapper[4756]: E0203 09:28:53.826843 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c\": container with ID starting with fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c not found: ID does not exist" containerID="fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.826888 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c"} err="failed to get container status \"fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c\": rpc error: code = NotFound desc = could not find container \"fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c\": container with ID starting with fd0c4c45801d712946ac8ca49b2be0081c9cefa2b1931422b9dba3f2bc28920c not found: ID does not exist" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.826915 4756 scope.go:117] "RemoveContainer" containerID="d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76" Feb 03 09:28:53 crc kubenswrapper[4756]: E0203 09:28:53.847704 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76\": container with ID starting with d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76 not found: ID does not exist" containerID="d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.847748 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76"} err="failed to get container status \"d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76\": rpc error: code = NotFound desc = could not find container \"d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76\": container with ID starting with d60afabee32ceba9e58c8d5d3731012a4dd3da4002a3ad7b840d92d046645f76 not found: ID does not exist" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.847775 4756 scope.go:117] "RemoveContainer" containerID="72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273" Feb 03 09:28:53 crc kubenswrapper[4756]: E0203 09:28:53.853790 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273\": container with ID starting with 72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273 not found: ID does not exist" containerID="72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.853841 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273"} err="failed to get container status \"72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273\": rpc error: code = NotFound desc = could not find container \"72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273\": container with ID starting with 72b1255c7dc8c9506cb9db5f9e5d0cde5e624b229bcd10fce2c39c648bfd7273 not found: ID does not exist" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.905302 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:28:53 crc kubenswrapper[4756]: I0203 09:28:53.972768 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.400509 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd03b9b6-ff6d-4eed-be6d-e08461b6c693","Type":"ContainerStarted","Data":"ef23578273bcfc98a4a44983f488407a2eaef76a5b1b615cb72f99ddab3850d9"} Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.406582 4756 generic.go:334] "Generic (PLEG): container finished" podID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" containerID="3ba8bbb10f06208e72abd6eb88537dc7a837d5ea2401f8b8305cfd5a7847149f" exitCode=0 Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.407366 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" event={"ID":"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae","Type":"ContainerDied","Data":"3ba8bbb10f06208e72abd6eb88537dc7a837d5ea2401f8b8305cfd5a7847149f"} Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.454474 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" event={"ID":"fb9890e6-45ec-47f2-969d-ffed8b40fb5c","Type":"ContainerStarted","Data":"8fea60c16d0173717be71e1375455613406263a7b257e07d285ac174b84d6127"} Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.454529 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" event={"ID":"fb9890e6-45ec-47f2-969d-ffed8b40fb5c","Type":"ContainerStarted","Data":"92e95bc13a65b86da37b2d83c10adf31e5bf7e548939c92189bed3ded00d134a"} Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.455670 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.455707 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.541182 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" podStartSLOduration=3.5411599689999997 podStartE2EDuration="3.541159969s" podCreationTimestamp="2026-02-03 09:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:54.498893927 +0000 UTC m=+1125.649361302" watchObservedRunningTime="2026-02-03 09:28:54.541159969 +0000 UTC m=+1125.691627344" Feb 03 09:28:54 crc kubenswrapper[4756]: I0203 09:28:54.594616 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:28:54 crc kubenswrapper[4756]: W0203 09:28:54.614344 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee2fadc1_7b6c_4d19_abaf_57ece18ec2c5.slice/crio-a803c8442edc19e01a55b0ab1fa906e01a6acc8073cb737d069588c1698c1482 WatchSource:0}: Error finding container a803c8442edc19e01a55b0ab1fa906e01a6acc8073cb737d069588c1698c1482: Status 404 returned error can't find the container with id a803c8442edc19e01a55b0ab1fa906e01a6acc8073cb737d069588c1698c1482 Feb 03 09:28:55 crc kubenswrapper[4756]: I0203 09:28:55.488974 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" event={"ID":"fb9890e6-45ec-47f2-969d-ffed8b40fb5c","Type":"ContainerStarted","Data":"a4ed74681d7ba4efc54983a31700306db3a4304eaf38418ef396ea4737d7bc16"} Feb 03 09:28:55 crc kubenswrapper[4756]: I0203 09:28:55.492286 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd03b9b6-ff6d-4eed-be6d-e08461b6c693","Type":"ContainerStarted","Data":"f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93"} Feb 03 09:28:55 crc kubenswrapper[4756]: I0203 09:28:55.494910 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" event={"ID":"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae","Type":"ContainerStarted","Data":"25f13ba113ff6857726f350d241462274b745ddd2f8b67087ade678372a050c4"} Feb 03 09:28:55 crc kubenswrapper[4756]: I0203 09:28:55.494960 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:28:55 crc kubenswrapper[4756]: I0203 09:28:55.497074 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerStarted","Data":"a803c8442edc19e01a55b0ab1fa906e01a6acc8073cb737d069588c1698c1482"} Feb 03 09:28:55 crc kubenswrapper[4756]: I0203 09:28:55.502120 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3dd59369-6dbb-4877-886c-76f19f87b90a","Type":"ContainerStarted","Data":"6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad"} Feb 03 09:28:55 crc kubenswrapper[4756]: I0203 09:28:55.515537 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" podStartSLOduration=4.5155169950000005 podStartE2EDuration="4.515516995s" podCreationTimestamp="2026-02-03 09:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:55.511370076 +0000 UTC m=+1126.661837451" watchObservedRunningTime="2026-02-03 09:28:55.515516995 +0000 UTC m=+1126.665984370" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.318751 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.576581 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cdf446f5c-8xnr7"] Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.576930 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cdf446f5c-8xnr7" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-httpd" containerID="cri-o://04ccca2b9393d6da33858a8f16a4af917eda96771ed9582ab23d1e05526c433b" gracePeriod=30 Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.577359 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cdf446f5c-8xnr7" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-api" containerID="cri-o://ab6dc282d7fe8bff4c4f1ba921f07b47a6680b5855d525ab5d7e63d20f3b24c6" gracePeriod=30 Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.585726 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6cdf446f5c-8xnr7" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.155:9696/\": EOF" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.628019 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8547fc4b5c-w4sjb"] Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.639942 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8547fc4b5c-w4sjb"] Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.640077 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.801215 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-internal-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.801282 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-httpd-config\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.801308 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-public-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.801387 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-combined-ca-bundle\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.801552 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2ct2\" (UniqueName: \"kubernetes.io/projected/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-kube-api-access-w2ct2\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.801643 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-ovndb-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.801678 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-config\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.903979 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-ovndb-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.904041 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-config\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.904115 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-internal-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.904158 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-httpd-config\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.904183 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-public-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.904219 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-combined-ca-bundle\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.904281 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2ct2\" (UniqueName: \"kubernetes.io/projected/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-kube-api-access-w2ct2\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.911260 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-config\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.911390 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-httpd-config\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.911815 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-ovndb-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.918127 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-public-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.918851 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-internal-tls-certs\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.931617 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-combined-ca-bundle\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.937535 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2ct2\" (UniqueName: \"kubernetes.io/projected/3a5ca842-66f7-4ff0-b8f7-afe33276b34b-kube-api-access-w2ct2\") pod \"neutron-8547fc4b5c-w4sjb\" (UID: \"3a5ca842-66f7-4ff0-b8f7-afe33276b34b\") " pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:57 crc kubenswrapper[4756]: I0203 09:28:57.972328 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.525970 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-f69df96f4-7db64"] Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.528107 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.531122 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.536912 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.566487 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f69df96f4-7db64"] Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.571141 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" event={"ID":"c34e66d1-0527-4f40-a676-1ac69c4d723f","Type":"ContainerStarted","Data":"582431fb5e5d0b27ea6a319344e46cb70ec4e421088c9f1e8ddf46371f514938"} Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.587673 4756 generic.go:334] "Generic (PLEG): container finished" podID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerID="04ccca2b9393d6da33858a8f16a4af917eda96771ed9582ab23d1e05526c433b" exitCode=0 Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.587775 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdf446f5c-8xnr7" event={"ID":"7fce406f-4aff-48f6-980a-64a65ede0c74","Type":"ContainerDied","Data":"04ccca2b9393d6da33858a8f16a4af917eda96771ed9582ab23d1e05526c433b"} Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.597013 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerStarted","Data":"bf36250fb9ab5d06431f72af20bbf4ee9068d16323f2d434636481097391a131"} Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.598904 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8547fc4b5c-w4sjb"] Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.612028 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3dd59369-6dbb-4877-886c-76f19f87b90a","Type":"ContainerStarted","Data":"f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa"} Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.627094 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-config-data-custom\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.627165 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-internal-tls-certs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.627379 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-combined-ca-bundle\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.627431 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-logs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.627536 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgs9l\" (UniqueName: \"kubernetes.io/projected/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-kube-api-access-kgs9l\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.627648 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-public-tls-certs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.627703 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-config-data\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.630787 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd03b9b6-ff6d-4eed-be6d-e08461b6c693","Type":"ContainerStarted","Data":"34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2"} Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.630994 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerName="cinder-api-log" containerID="cri-o://f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93" gracePeriod=30 Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.631313 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.631363 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerName="cinder-api" containerID="cri-o://34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2" gracePeriod=30 Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.648363 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b68864ff-4znh2" event={"ID":"8b049610-c8e6-40bd-98cb-138f42f1c35f","Type":"ContainerStarted","Data":"2bac71edb1f19f119278e8ad0ba9941a2b652f2baacdb516d8d362c1b0949afe"} Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.674311 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.639779646 podStartE2EDuration="7.674280153s" podCreationTimestamp="2026-02-03 09:28:51 +0000 UTC" firstStartedPulling="2026-02-03 09:28:52.609358457 +0000 UTC m=+1123.759825832" lastFinishedPulling="2026-02-03 09:28:53.643858964 +0000 UTC m=+1124.794326339" observedRunningTime="2026-02-03 09:28:58.638554976 +0000 UTC m=+1129.789022361" watchObservedRunningTime="2026-02-03 09:28:58.674280153 +0000 UTC m=+1129.824747528" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.682267 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.682242582 podStartE2EDuration="7.682242582s" podCreationTimestamp="2026-02-03 09:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:28:58.662922388 +0000 UTC m=+1129.813389773" watchObservedRunningTime="2026-02-03 09:28:58.682242582 +0000 UTC m=+1129.832709957" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.730507 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-config-data-custom\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.730577 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-internal-tls-certs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.730668 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-combined-ca-bundle\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.730703 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-logs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.730760 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgs9l\" (UniqueName: \"kubernetes.io/projected/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-kube-api-access-kgs9l\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.730822 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-public-tls-certs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.730880 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-config-data\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.744931 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-logs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.745334 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-internal-tls-certs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.745685 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-config-data\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.751745 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-config-data-custom\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.756491 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-public-tls-certs\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.764664 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-combined-ca-bundle\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.770514 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgs9l\" (UniqueName: \"kubernetes.io/projected/ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f-kube-api-access-kgs9l\") pod \"barbican-api-f69df96f4-7db64\" (UID: \"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f\") " pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:58 crc kubenswrapper[4756]: I0203 09:28:58.886862 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.446270 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6cdf446f5c-8xnr7" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.155:9696/\": dial tcp 10.217.0.155:9696: connect: connection refused" Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.476982 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f69df96f4-7db64"] Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.750597 4756 generic.go:334] "Generic (PLEG): container finished" podID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerID="f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93" exitCode=143 Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.750896 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd03b9b6-ff6d-4eed-be6d-e08461b6c693","Type":"ContainerDied","Data":"f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93"} Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.785945 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69df96f4-7db64" event={"ID":"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f","Type":"ContainerStarted","Data":"4104aab4f18e72d0bb8fcd7c905a0848c59ea356d568d061ffdfbc16083a8b24"} Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.791542 4756 generic.go:334] "Generic (PLEG): container finished" podID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerID="a2e41bbc539623bf400c55db9e7628d7435096b137b557a7fb505514f8eec9c5" exitCode=137 Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.791572 4756 generic.go:334] "Generic (PLEG): container finished" podID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerID="5f5ead35d19b5efc89edc6b946b5e705dec6a56c9da7887aff9733cdc2c84350" exitCode=137 Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.791611 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bb66dd6bf-zpxb2" event={"ID":"a9c7b8f6-aef0-404b-b0df-6f716518aa01","Type":"ContainerDied","Data":"a2e41bbc539623bf400c55db9e7628d7435096b137b557a7fb505514f8eec9c5"} Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.791670 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bb66dd6bf-zpxb2" event={"ID":"a9c7b8f6-aef0-404b-b0df-6f716518aa01","Type":"ContainerDied","Data":"5f5ead35d19b5efc89edc6b946b5e705dec6a56c9da7887aff9733cdc2c84350"} Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.794075 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68b68864ff-4znh2" event={"ID":"8b049610-c8e6-40bd-98cb-138f42f1c35f","Type":"ContainerStarted","Data":"dab72ee9f53f4ca42c278b52a7a75b64c4b5516a2c40456039a6ad10d60f0c44"} Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.801267 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" event={"ID":"c34e66d1-0527-4f40-a676-1ac69c4d723f","Type":"ContainerStarted","Data":"2bc73503785f858e8129905ea76fc040b293c254238d2cd1c2f4abf06587632b"} Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.803627 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8547fc4b5c-w4sjb" event={"ID":"3a5ca842-66f7-4ff0-b8f7-afe33276b34b","Type":"ContainerStarted","Data":"054dee14d46fee93bac0b5f68b94355e07773212c26f21c26988abe1afe5c7b8"} Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.803650 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8547fc4b5c-w4sjb" event={"ID":"3a5ca842-66f7-4ff0-b8f7-afe33276b34b","Type":"ContainerStarted","Data":"0ef25b83e3626c10e37adcbad965f3aaeb1454cafd262195e33eed215be51e19"} Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.829104 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-68b68864ff-4znh2" podStartSLOduration=6.071681638 podStartE2EDuration="8.829088293s" podCreationTimestamp="2026-02-03 09:28:51 +0000 UTC" firstStartedPulling="2026-02-03 09:28:52.853771522 +0000 UTC m=+1124.004238897" lastFinishedPulling="2026-02-03 09:28:55.611178177 +0000 UTC m=+1126.761645552" observedRunningTime="2026-02-03 09:28:59.81909593 +0000 UTC m=+1130.969563305" watchObservedRunningTime="2026-02-03 09:28:59.829088293 +0000 UTC m=+1130.979555668" Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.843193 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.850082 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5948f79bd-cl5cq" podStartSLOduration=6.251910935 podStartE2EDuration="8.850065089s" podCreationTimestamp="2026-02-03 09:28:51 +0000 UTC" firstStartedPulling="2026-02-03 09:28:53.030571802 +0000 UTC m=+1124.181039177" lastFinishedPulling="2026-02-03 09:28:55.628725966 +0000 UTC m=+1126.779193331" observedRunningTime="2026-02-03 09:28:59.849429379 +0000 UTC m=+1130.999896754" watchObservedRunningTime="2026-02-03 09:28:59.850065089 +0000 UTC m=+1131.000532464" Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.971145 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9c7b8f6-aef0-404b-b0df-6f716518aa01-logs\") pod \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.971254 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-scripts\") pod \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.971369 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqdpj\" (UniqueName: \"kubernetes.io/projected/a9c7b8f6-aef0-404b-b0df-6f716518aa01-kube-api-access-wqdpj\") pod \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.971437 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-config-data\") pod \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.971484 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9c7b8f6-aef0-404b-b0df-6f716518aa01-horizon-secret-key\") pod \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\" (UID: \"a9c7b8f6-aef0-404b-b0df-6f716518aa01\") " Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.972978 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9c7b8f6-aef0-404b-b0df-6f716518aa01-logs" (OuterVolumeSpecName: "logs") pod "a9c7b8f6-aef0-404b-b0df-6f716518aa01" (UID: "a9c7b8f6-aef0-404b-b0df-6f716518aa01"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.983376 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9c7b8f6-aef0-404b-b0df-6f716518aa01-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a9c7b8f6-aef0-404b-b0df-6f716518aa01" (UID: "a9c7b8f6-aef0-404b-b0df-6f716518aa01"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:28:59 crc kubenswrapper[4756]: I0203 09:28:59.990838 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9c7b8f6-aef0-404b-b0df-6f716518aa01-kube-api-access-wqdpj" (OuterVolumeSpecName: "kube-api-access-wqdpj") pod "a9c7b8f6-aef0-404b-b0df-6f716518aa01" (UID: "a9c7b8f6-aef0-404b-b0df-6f716518aa01"). InnerVolumeSpecName "kube-api-access-wqdpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.073408 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqdpj\" (UniqueName: \"kubernetes.io/projected/a9c7b8f6-aef0-404b-b0df-6f716518aa01-kube-api-access-wqdpj\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.073436 4756 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a9c7b8f6-aef0-404b-b0df-6f716518aa01-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.073456 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9c7b8f6-aef0-404b-b0df-6f716518aa01-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.083121 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-scripts" (OuterVolumeSpecName: "scripts") pod "a9c7b8f6-aef0-404b-b0df-6f716518aa01" (UID: "a9c7b8f6-aef0-404b-b0df-6f716518aa01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.085399 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-config-data" (OuterVolumeSpecName: "config-data") pod "a9c7b8f6-aef0-404b-b0df-6f716518aa01" (UID: "a9c7b8f6-aef0-404b-b0df-6f716518aa01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.177335 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.177361 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a9c7b8f6-aef0-404b-b0df-6f716518aa01-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.814417 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerStarted","Data":"64793391177c891ddb316dea511f27d059e292e92e021d3f67b7b74a6e7449f5"} Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.816688 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8547fc4b5c-w4sjb" event={"ID":"3a5ca842-66f7-4ff0-b8f7-afe33276b34b","Type":"ContainerStarted","Data":"2927c8b6a59f21342d1339dd9cb2b8269b03c1a3ab84f587c0aa78a153454b11"} Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.816833 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.818728 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69df96f4-7db64" event={"ID":"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f","Type":"ContainerStarted","Data":"0f5f42721d69ccb04403fdf5bc2698a12cac9d2e4e68bd3dda8c3dec437d2bc6"} Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.818760 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f69df96f4-7db64" event={"ID":"ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f","Type":"ContainerStarted","Data":"13b4375841d094256a3fbc3094af7bcff0970a7ddc6003a31b2a784be207f720"} Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.818914 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.820657 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bb66dd6bf-zpxb2" event={"ID":"a9c7b8f6-aef0-404b-b0df-6f716518aa01","Type":"ContainerDied","Data":"0c961a97c561677505b4b8e97b5c972c226359347da9f02f465a6247cb25a600"} Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.820705 4756 scope.go:117] "RemoveContainer" containerID="a2e41bbc539623bf400c55db9e7628d7435096b137b557a7fb505514f8eec9c5" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.820764 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bb66dd6bf-zpxb2" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.845135 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8547fc4b5c-w4sjb" podStartSLOduration=3.845114372 podStartE2EDuration="3.845114372s" podCreationTimestamp="2026-02-03 09:28:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:00.839138565 +0000 UTC m=+1131.989605960" watchObservedRunningTime="2026-02-03 09:29:00.845114372 +0000 UTC m=+1131.995581747" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.869462 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-f69df96f4-7db64" podStartSLOduration=2.869431122 podStartE2EDuration="2.869431122s" podCreationTimestamp="2026-02-03 09:28:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:00.864943401 +0000 UTC m=+1132.015410776" watchObservedRunningTime="2026-02-03 09:29:00.869431122 +0000 UTC m=+1132.019898497" Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.902674 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bb66dd6bf-zpxb2"] Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.913726 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bb66dd6bf-zpxb2"] Feb 03 09:29:00 crc kubenswrapper[4756]: I0203 09:29:00.968697 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:29:01 crc kubenswrapper[4756]: I0203 09:29:01.056010 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:29:01 crc kubenswrapper[4756]: I0203 09:29:01.071512 4756 scope.go:117] "RemoveContainer" containerID="5f5ead35d19b5efc89edc6b946b5e705dec6a56c9da7887aff9733cdc2c84350" Feb 03 09:29:01 crc kubenswrapper[4756]: I0203 09:29:01.626148 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" path="/var/lib/kubelet/pods/a9c7b8f6-aef0-404b-b0df-6f716518aa01/volumes" Feb 03 09:29:01 crc kubenswrapper[4756]: I0203 09:29:01.830916 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerStarted","Data":"6cd0d88ad9342418a15d0df2f356663f9a991e202daf83f6b323fd429bb10661"} Feb 03 09:29:01 crc kubenswrapper[4756]: I0203 09:29:01.831156 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:29:01 crc kubenswrapper[4756]: I0203 09:29:01.999080 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 03 09:29:02 crc kubenswrapper[4756]: I0203 09:29:02.230045 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 03 09:29:02 crc kubenswrapper[4756]: I0203 09:29:02.499622 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:29:02 crc kubenswrapper[4756]: I0203 09:29:02.555861 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-mm4zx"] Feb 03 09:29:02 crc kubenswrapper[4756]: I0203 09:29:02.556087 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" podUID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" containerName="dnsmasq-dns" containerID="cri-o://2b206a6e16728314c6e93f23fc70347b7c70b457f411ac981811d128d4b7b258" gracePeriod=10 Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.146407 4756 generic.go:334] "Generic (PLEG): container finished" podID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" containerID="2b206a6e16728314c6e93f23fc70347b7c70b457f411ac981811d128d4b7b258" exitCode=0 Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.146477 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" event={"ID":"3beb75ec-64c9-413b-9efe-63a05ddb0bab","Type":"ContainerDied","Data":"2b206a6e16728314c6e93f23fc70347b7c70b457f411ac981811d128d4b7b258"} Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.150268 4756 generic.go:334] "Generic (PLEG): container finished" podID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerID="ab6dc282d7fe8bff4c4f1ba921f07b47a6680b5855d525ab5d7e63d20f3b24c6" exitCode=0 Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.151182 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdf446f5c-8xnr7" event={"ID":"7fce406f-4aff-48f6-980a-64a65ede0c74","Type":"ContainerDied","Data":"ab6dc282d7fe8bff4c4f1ba921f07b47a6680b5855d525ab5d7e63d20f3b24c6"} Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.279646 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.513737 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.684130 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-public-tls-certs\") pod \"7fce406f-4aff-48f6-980a-64a65ede0c74\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.684197 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x657q\" (UniqueName: \"kubernetes.io/projected/7fce406f-4aff-48f6-980a-64a65ede0c74-kube-api-access-x657q\") pod \"7fce406f-4aff-48f6-980a-64a65ede0c74\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.684357 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-combined-ca-bundle\") pod \"7fce406f-4aff-48f6-980a-64a65ede0c74\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.684400 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-ovndb-tls-certs\") pod \"7fce406f-4aff-48f6-980a-64a65ede0c74\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.684479 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-httpd-config\") pod \"7fce406f-4aff-48f6-980a-64a65ede0c74\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.684548 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-internal-tls-certs\") pod \"7fce406f-4aff-48f6-980a-64a65ede0c74\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.684603 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-config\") pod \"7fce406f-4aff-48f6-980a-64a65ede0c74\" (UID: \"7fce406f-4aff-48f6-980a-64a65ede0c74\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.693553 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fce406f-4aff-48f6-980a-64a65ede0c74-kube-api-access-x657q" (OuterVolumeSpecName: "kube-api-access-x657q") pod "7fce406f-4aff-48f6-980a-64a65ede0c74" (UID: "7fce406f-4aff-48f6-980a-64a65ede0c74"). InnerVolumeSpecName "kube-api-access-x657q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.696022 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7fce406f-4aff-48f6-980a-64a65ede0c74" (UID: "7fce406f-4aff-48f6-980a-64a65ede0c74"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.752607 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-config" (OuterVolumeSpecName: "config") pod "7fce406f-4aff-48f6-980a-64a65ede0c74" (UID: "7fce406f-4aff-48f6-980a-64a65ede0c74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.761423 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fce406f-4aff-48f6-980a-64a65ede0c74" (UID: "7fce406f-4aff-48f6-980a-64a65ede0c74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.781248 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "7fce406f-4aff-48f6-980a-64a65ede0c74" (UID: "7fce406f-4aff-48f6-980a-64a65ede0c74"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.786369 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.786388 4756 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.786398 4756 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.786406 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.786414 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x657q\" (UniqueName: \"kubernetes.io/projected/7fce406f-4aff-48f6-980a-64a65ede0c74-kube-api-access-x657q\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.794670 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7fce406f-4aff-48f6-980a-64a65ede0c74" (UID: "7fce406f-4aff-48f6-980a-64a65ede0c74"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.794706 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7fce406f-4aff-48f6-980a-64a65ede0c74" (UID: "7fce406f-4aff-48f6-980a-64a65ede0c74"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.873538 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.887555 4756 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.887595 4756 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fce406f-4aff-48f6-980a-64a65ede0c74-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.890985 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6d95f7b7bb-r6fd8" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.954618 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-546db499dd-c5f8j"] Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.954841 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-546db499dd-c5f8j" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon-log" containerID="cri-o://798d8b64be1bedcbb904e87c52b8e23badc5c7864eb3968e0e041473058e8551" gracePeriod=30 Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.955202 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-546db499dd-c5f8j" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" containerID="cri-o://4f5b4620a588f33eab38c31e48cf91bf8ea74628a6ed552028a79834e5626fdb" gracePeriod=30 Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.971886 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-546db499dd-c5f8j" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.991005 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-sb\") pod \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.991149 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxtzc\" (UniqueName: \"kubernetes.io/projected/3beb75ec-64c9-413b-9efe-63a05ddb0bab-kube-api-access-nxtzc\") pod \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.991250 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-nb\") pod \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.991289 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-config\") pod \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.991315 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-svc\") pod \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.991346 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-swift-storage-0\") pod \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\" (UID: \"3beb75ec-64c9-413b-9efe-63a05ddb0bab\") " Feb 03 09:29:03 crc kubenswrapper[4756]: I0203 09:29:03.998650 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3beb75ec-64c9-413b-9efe-63a05ddb0bab-kube-api-access-nxtzc" (OuterVolumeSpecName: "kube-api-access-nxtzc") pod "3beb75ec-64c9-413b-9efe-63a05ddb0bab" (UID: "3beb75ec-64c9-413b-9efe-63a05ddb0bab"). InnerVolumeSpecName "kube-api-access-nxtzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.054158 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3beb75ec-64c9-413b-9efe-63a05ddb0bab" (UID: "3beb75ec-64c9-413b-9efe-63a05ddb0bab"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.065702 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3beb75ec-64c9-413b-9efe-63a05ddb0bab" (UID: "3beb75ec-64c9-413b-9efe-63a05ddb0bab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.069904 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-config" (OuterVolumeSpecName: "config") pod "3beb75ec-64c9-413b-9efe-63a05ddb0bab" (UID: "3beb75ec-64c9-413b-9efe-63a05ddb0bab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.079167 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3beb75ec-64c9-413b-9efe-63a05ddb0bab" (UID: "3beb75ec-64c9-413b-9efe-63a05ddb0bab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.079625 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3beb75ec-64c9-413b-9efe-63a05ddb0bab" (UID: "3beb75ec-64c9-413b-9efe-63a05ddb0bab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.095104 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.095144 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.095190 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.095203 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.095212 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3beb75ec-64c9-413b-9efe-63a05ddb0bab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.095256 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxtzc\" (UniqueName: \"kubernetes.io/projected/3beb75ec-64c9-413b-9efe-63a05ddb0bab-kube-api-access-nxtzc\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.162249 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" event={"ID":"3beb75ec-64c9-413b-9efe-63a05ddb0bab","Type":"ContainerDied","Data":"5489ed7f3a51c5e1ff968562b05333d4b9957c9e8f954ad961b1468a254c34db"} Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.162309 4756 scope.go:117] "RemoveContainer" containerID="2b206a6e16728314c6e93f23fc70347b7c70b457f411ac981811d128d4b7b258" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.162308 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-mm4zx" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.165244 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdf446f5c-8xnr7" event={"ID":"7fce406f-4aff-48f6-980a-64a65ede0c74","Type":"ContainerDied","Data":"0cb64b5542a53c0cbae8d0bccb77ca32ef6e6ac36f7b885c4b35edac0fdb6a24"} Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.165267 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cdf446f5c-8xnr7" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.167980 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerName="cinder-scheduler" containerID="cri-o://6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad" gracePeriod=30 Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.168235 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerName="probe" containerID="cri-o://f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa" gracePeriod=30 Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.168502 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerStarted","Data":"420fee0f8ce0be289bb2f7da9a5eadb9a5bf8f3a7912485e9a08f7b1a3453a04"} Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.169169 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.198660 4756 scope.go:117] "RemoveContainer" containerID="7c2cfc72d439839f4aa8727963f7dd9a27d3f3e373f4d58c6b4ded60402f5db1" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.217079 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.507931347 podStartE2EDuration="11.217060067s" podCreationTimestamp="2026-02-03 09:28:53 +0000 UTC" firstStartedPulling="2026-02-03 09:28:54.616779854 +0000 UTC m=+1125.767247239" lastFinishedPulling="2026-02-03 09:29:03.325908594 +0000 UTC m=+1134.476375959" observedRunningTime="2026-02-03 09:29:04.211169893 +0000 UTC m=+1135.361637268" watchObservedRunningTime="2026-02-03 09:29:04.217060067 +0000 UTC m=+1135.367527442" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.239189 4756 scope.go:117] "RemoveContainer" containerID="04ccca2b9393d6da33858a8f16a4af917eda96771ed9582ab23d1e05526c433b" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.240957 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-mm4zx"] Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.251349 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-mm4zx"] Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.264236 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cdf446f5c-8xnr7"] Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.265906 4756 scope.go:117] "RemoveContainer" containerID="ab6dc282d7fe8bff4c4f1ba921f07b47a6680b5855d525ab5d7e63d20f3b24c6" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.273219 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6cdf446f5c-8xnr7"] Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.827587 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:29:04 crc kubenswrapper[4756]: I0203 09:29:04.941135 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:29:05 crc kubenswrapper[4756]: I0203 09:29:05.179496 4756 generic.go:334] "Generic (PLEG): container finished" podID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerID="f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa" exitCode=0 Feb 03 09:29:05 crc kubenswrapper[4756]: I0203 09:29:05.179584 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3dd59369-6dbb-4877-886c-76f19f87b90a","Type":"ContainerDied","Data":"f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa"} Feb 03 09:29:05 crc kubenswrapper[4756]: I0203 09:29:05.626155 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" path="/var/lib/kubelet/pods/3beb75ec-64c9-413b-9efe-63a05ddb0bab/volumes" Feb 03 09:29:05 crc kubenswrapper[4756]: I0203 09:29:05.626934 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" path="/var/lib/kubelet/pods/7fce406f-4aff-48f6-980a-64a65ede0c74/volumes" Feb 03 09:29:07 crc kubenswrapper[4756]: I0203 09:29:07.095061 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-546db499dd-c5f8j" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:43828->10.217.0.148:8443: read: connection reset by peer" Feb 03 09:29:07 crc kubenswrapper[4756]: I0203 09:29:07.202858 4756 generic.go:334] "Generic (PLEG): container finished" podID="e45bd2fd-140c-409a-9592-252f4723354e" containerID="4f5b4620a588f33eab38c31e48cf91bf8ea74628a6ed552028a79834e5626fdb" exitCode=0 Feb 03 09:29:07 crc kubenswrapper[4756]: I0203 09:29:07.202916 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-546db499dd-c5f8j" event={"ID":"e45bd2fd-140c-409a-9592-252f4723354e","Type":"ContainerDied","Data":"4f5b4620a588f33eab38c31e48cf91bf8ea74628a6ed552028a79834e5626fdb"} Feb 03 09:29:07 crc kubenswrapper[4756]: I0203 09:29:07.984428 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-ccf85886-5xql4" Feb 03 09:29:07 crc kubenswrapper[4756]: I0203 09:29:07.986699 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-ccf85886-5xql4" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.166976 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-546db499dd-c5f8j" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.248613 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6f5fdf688d-jsswq"] Feb 03 09:29:08 crc kubenswrapper[4756]: E0203 09:29:08.249131 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-httpd" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249155 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-httpd" Feb 03 09:29:08 crc kubenswrapper[4756]: E0203 09:29:08.249166 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerName="horizon" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249174 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerName="horizon" Feb 03 09:29:08 crc kubenswrapper[4756]: E0203 09:29:08.249195 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" containerName="dnsmasq-dns" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249204 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" containerName="dnsmasq-dns" Feb 03 09:29:08 crc kubenswrapper[4756]: E0203 09:29:08.249217 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" containerName="init" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249224 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" containerName="init" Feb 03 09:29:08 crc kubenswrapper[4756]: E0203 09:29:08.249248 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerName="horizon-log" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249256 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerName="horizon-log" Feb 03 09:29:08 crc kubenswrapper[4756]: E0203 09:29:08.249273 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-api" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249281 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-api" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249491 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerName="horizon" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249515 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9c7b8f6-aef0-404b-b0df-6f716518aa01" containerName="horizon-log" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249535 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-httpd" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249548 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3beb75ec-64c9-413b-9efe-63a05ddb0bab" containerName="dnsmasq-dns" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.249560 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fce406f-4aff-48f6-980a-64a65ede0c74" containerName="neutron-api" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.250591 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.289358 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6f5fdf688d-jsswq"] Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.385404 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-config-data\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.385512 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-internal-tls-certs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.385570 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-public-tls-certs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.385674 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-scripts\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.385710 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-combined-ca-bundle\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.385869 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fgmh\" (UniqueName: \"kubernetes.io/projected/5446426b-adaf-4af7-97f8-b8b2d798d364-kube-api-access-6fgmh\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.385931 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5446426b-adaf-4af7-97f8-b8b2d798d364-logs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.487801 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-scripts\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.487860 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-combined-ca-bundle\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.487918 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fgmh\" (UniqueName: \"kubernetes.io/projected/5446426b-adaf-4af7-97f8-b8b2d798d364-kube-api-access-6fgmh\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.487946 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5446426b-adaf-4af7-97f8-b8b2d798d364-logs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.488015 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-config-data\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.488046 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-internal-tls-certs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.488073 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-public-tls-certs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.489336 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5446426b-adaf-4af7-97f8-b8b2d798d364-logs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.495744 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-config-data\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.496556 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-combined-ca-bundle\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.496765 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-scripts\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.497837 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-public-tls-certs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.501490 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5446426b-adaf-4af7-97f8-b8b2d798d364-internal-tls-certs\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.540672 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fgmh\" (UniqueName: \"kubernetes.io/projected/5446426b-adaf-4af7-97f8-b8b2d798d364-kube-api-access-6fgmh\") pod \"placement-6f5fdf688d-jsswq\" (UID: \"5446426b-adaf-4af7-97f8-b8b2d798d364\") " pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.584776 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.786648 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.899275 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data-custom\") pod \"3dd59369-6dbb-4877-886c-76f19f87b90a\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.899382 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-scripts\") pod \"3dd59369-6dbb-4877-886c-76f19f87b90a\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.899536 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data\") pod \"3dd59369-6dbb-4877-886c-76f19f87b90a\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.900299 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3dd59369-6dbb-4877-886c-76f19f87b90a-etc-machine-id\") pod \"3dd59369-6dbb-4877-886c-76f19f87b90a\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.900356 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-combined-ca-bundle\") pod \"3dd59369-6dbb-4877-886c-76f19f87b90a\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.900394 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj44g\" (UniqueName: \"kubernetes.io/projected/3dd59369-6dbb-4877-886c-76f19f87b90a-kube-api-access-jj44g\") pod \"3dd59369-6dbb-4877-886c-76f19f87b90a\" (UID: \"3dd59369-6dbb-4877-886c-76f19f87b90a\") " Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.902420 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3dd59369-6dbb-4877-886c-76f19f87b90a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3dd59369-6dbb-4877-886c-76f19f87b90a" (UID: "3dd59369-6dbb-4877-886c-76f19f87b90a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.906932 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-scripts" (OuterVolumeSpecName: "scripts") pod "3dd59369-6dbb-4877-886c-76f19f87b90a" (UID: "3dd59369-6dbb-4877-886c-76f19f87b90a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.909133 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3dd59369-6dbb-4877-886c-76f19f87b90a" (UID: "3dd59369-6dbb-4877-886c-76f19f87b90a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.911010 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd59369-6dbb-4877-886c-76f19f87b90a-kube-api-access-jj44g" (OuterVolumeSpecName: "kube-api-access-jj44g") pod "3dd59369-6dbb-4877-886c-76f19f87b90a" (UID: "3dd59369-6dbb-4877-886c-76f19f87b90a"). InnerVolumeSpecName "kube-api-access-jj44g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:08 crc kubenswrapper[4756]: I0203 09:29:08.976962 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dd59369-6dbb-4877-886c-76f19f87b90a" (UID: "3dd59369-6dbb-4877-886c-76f19f87b90a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.002661 4756 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3dd59369-6dbb-4877-886c-76f19f87b90a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.002961 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.002971 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj44g\" (UniqueName: \"kubernetes.io/projected/3dd59369-6dbb-4877-886c-76f19f87b90a-kube-api-access-jj44g\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.002983 4756 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.002991 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.013718 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data" (OuterVolumeSpecName: "config-data") pod "3dd59369-6dbb-4877-886c-76f19f87b90a" (UID: "3dd59369-6dbb-4877-886c-76f19f87b90a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.104235 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd59369-6dbb-4877-886c-76f19f87b90a-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.223277 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6f5fdf688d-jsswq"] Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.246924 4756 generic.go:334] "Generic (PLEG): container finished" podID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerID="6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad" exitCode=0 Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.246999 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3dd59369-6dbb-4877-886c-76f19f87b90a","Type":"ContainerDied","Data":"6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad"} Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.247065 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3dd59369-6dbb-4877-886c-76f19f87b90a","Type":"ContainerDied","Data":"bbd3136bd4711b71de4852098f921d420d4f67898c0a25dff7afb002e3003bb6"} Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.247092 4756 scope.go:117] "RemoveContainer" containerID="f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.247401 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.282579 4756 scope.go:117] "RemoveContainer" containerID="6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.334785 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.374310 4756 scope.go:117] "RemoveContainer" containerID="f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.374425 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.374471 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:29:09 crc kubenswrapper[4756]: E0203 09:29:09.376131 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerName="probe" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.376154 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerName="probe" Feb 03 09:29:09 crc kubenswrapper[4756]: E0203 09:29:09.376167 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerName="cinder-scheduler" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.376174 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerName="cinder-scheduler" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.376367 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerName="probe" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.376382 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" containerName="cinder-scheduler" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.377353 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.380123 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 03 09:29:09 crc kubenswrapper[4756]: E0203 09:29:09.381776 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa\": container with ID starting with f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa not found: ID does not exist" containerID="f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.381816 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa"} err="failed to get container status \"f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa\": rpc error: code = NotFound desc = could not find container \"f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa\": container with ID starting with f5b6445e9a81d7d87105f0cf67f97b334fcc909015f0f5f062f3929556cdb3aa not found: ID does not exist" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.381842 4756 scope.go:117] "RemoveContainer" containerID="6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad" Feb 03 09:29:09 crc kubenswrapper[4756]: E0203 09:29:09.382174 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad\": container with ID starting with 6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad not found: ID does not exist" containerID="6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.382209 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad"} err="failed to get container status \"6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad\": rpc error: code = NotFound desc = could not find container \"6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad\": container with ID starting with 6a5da98192fed6913e29821736dd715c200c328e667aa20500e0112e59d26fad not found: ID does not exist" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.392471 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.519649 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkvd7\" (UniqueName: \"kubernetes.io/projected/5abc83aa-9e60-4884-9633-d151104a151a-kube-api-access-zkvd7\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.519710 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.519887 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.519926 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-scripts\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.519986 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5abc83aa-9e60-4884-9633-d151104a151a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.520025 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-config-data\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.624952 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkvd7\" (UniqueName: \"kubernetes.io/projected/5abc83aa-9e60-4884-9633-d151104a151a-kube-api-access-zkvd7\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.625325 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.625483 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.625509 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-scripts\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.625561 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5abc83aa-9e60-4884-9633-d151104a151a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.625601 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-config-data\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.631517 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5abc83aa-9e60-4884-9633-d151104a151a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.632738 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-config-data\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.632751 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dd59369-6dbb-4877-886c-76f19f87b90a" path="/var/lib/kubelet/pods/3dd59369-6dbb-4877-886c-76f19f87b90a/volumes" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.637736 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.642985 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.648552 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-scripts\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.648896 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abc83aa-9e60-4884-9633-d151104a151a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.655162 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkvd7\" (UniqueName: \"kubernetes.io/projected/5abc83aa-9e60-4884-9633-d151104a151a-kube-api-access-zkvd7\") pod \"cinder-scheduler-0\" (UID: \"5abc83aa-9e60-4884-9633-d151104a151a\") " pod="openstack/cinder-scheduler-0" Feb 03 09:29:09 crc kubenswrapper[4756]: I0203 09:29:09.814684 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.259199 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f5fdf688d-jsswq" event={"ID":"5446426b-adaf-4af7-97f8-b8b2d798d364","Type":"ContainerStarted","Data":"6b2c058d4ba3cdffbdd7a79a0cc7614c0f2c8d532ddd248bdd7e1ac089f1f0e8"} Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.259563 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f5fdf688d-jsswq" event={"ID":"5446426b-adaf-4af7-97f8-b8b2d798d364","Type":"ContainerStarted","Data":"0664ff3370147c9d5d252251f37e243bcafc28cd370c35c50ba3e8e52e736c11"} Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.259581 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f5fdf688d-jsswq" event={"ID":"5446426b-adaf-4af7-97f8-b8b2d798d364","Type":"ContainerStarted","Data":"441e7e6348504c182a26db2345ca175357ac8781ff9d56e4614adc22b813214c"} Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.260239 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.260285 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.301619 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6f5fdf688d-jsswq" podStartSLOduration=2.301597926 podStartE2EDuration="2.301597926s" podCreationTimestamp="2026-02-03 09:29:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:10.287905378 +0000 UTC m=+1141.438372753" watchObservedRunningTime="2026-02-03 09:29:10.301597926 +0000 UTC m=+1141.452065301" Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.331729 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 03 09:29:10 crc kubenswrapper[4756]: W0203 09:29:10.337504 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5abc83aa_9e60_4884_9633_d151104a151a.slice/crio-4741d95d4af7cc316473c3c78a08be00185bc5f2a3382d857e668a4eb20e496c WatchSource:0}: Error finding container 4741d95d4af7cc316473c3c78a08be00185bc5f2a3382d857e668a4eb20e496c: Status 404 returned error can't find the container with id 4741d95d4af7cc316473c3c78a08be00185bc5f2a3382d857e668a4eb20e496c Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.467190 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.639123 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.741105 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f69df96f4-7db64" Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.792420 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b7c55c5fb-l7j2x"] Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.792690 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api-log" containerID="cri-o://8fea60c16d0173717be71e1375455613406263a7b257e07d285ac174b84d6127" gracePeriod=30 Feb 03 09:29:10 crc kubenswrapper[4756]: I0203 09:29:10.793273 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api" containerID="cri-o://a4ed74681d7ba4efc54983a31700306db3a4304eaf38418ef396ea4737d7bc16" gracePeriod=30 Feb 03 09:29:11 crc kubenswrapper[4756]: I0203 09:29:11.277565 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" event={"ID":"fb9890e6-45ec-47f2-969d-ffed8b40fb5c","Type":"ContainerDied","Data":"8fea60c16d0173717be71e1375455613406263a7b257e07d285ac174b84d6127"} Feb 03 09:29:11 crc kubenswrapper[4756]: I0203 09:29:11.277575 4756 generic.go:334] "Generic (PLEG): container finished" podID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerID="8fea60c16d0173717be71e1375455613406263a7b257e07d285ac174b84d6127" exitCode=143 Feb 03 09:29:11 crc kubenswrapper[4756]: I0203 09:29:11.279978 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5abc83aa-9e60-4884-9633-d151104a151a","Type":"ContainerStarted","Data":"820cbbfb5a3032556c2acd83824e1bf6e84a71035e55a339e0f8ecbca5307f0a"} Feb 03 09:29:11 crc kubenswrapper[4756]: I0203 09:29:11.280514 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5abc83aa-9e60-4884-9633-d151104a151a","Type":"ContainerStarted","Data":"4741d95d4af7cc316473c3c78a08be00185bc5f2a3382d857e668a4eb20e496c"} Feb 03 09:29:12 crc kubenswrapper[4756]: I0203 09:29:12.291668 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5abc83aa-9e60-4884-9633-d151104a151a","Type":"ContainerStarted","Data":"59a79f5afac73e013c18505a34f52893134874557e20fc0423802c8b5b368ef0"} Feb 03 09:29:12 crc kubenswrapper[4756]: I0203 09:29:12.882431 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-765cf46959-4rcst" Feb 03 09:29:12 crc kubenswrapper[4756]: I0203 09:29:12.912819 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.912787357 podStartE2EDuration="3.912787357s" podCreationTimestamp="2026-02-03 09:29:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:12.331271579 +0000 UTC m=+1143.481738954" watchObservedRunningTime="2026-02-03 09:29:12.912787357 +0000 UTC m=+1144.063254772" Feb 03 09:29:13 crc kubenswrapper[4756]: I0203 09:29:13.566480 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:29:13 crc kubenswrapper[4756]: I0203 09:29:13.566552 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:29:13 crc kubenswrapper[4756]: I0203 09:29:13.566607 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:29:13 crc kubenswrapper[4756]: I0203 09:29:13.567349 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0da902e4b7c741d00f45d02b231b9b641521e6c5b54725cc76fd568b0058caf7"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:29:13 crc kubenswrapper[4756]: I0203 09:29:13.567412 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://0da902e4b7c741d00f45d02b231b9b641521e6c5b54725cc76fd568b0058caf7" gracePeriod=600 Feb 03 09:29:13 crc kubenswrapper[4756]: I0203 09:29:13.969238 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:55946->10.217.0.163:9311: read: connection reset by peer" Feb 03 09:29:13 crc kubenswrapper[4756]: I0203 09:29:13.969297 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:55952->10.217.0.163:9311: read: connection reset by peer" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.315730 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="0da902e4b7c741d00f45d02b231b9b641521e6c5b54725cc76fd568b0058caf7" exitCode=0 Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.315780 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"0da902e4b7c741d00f45d02b231b9b641521e6c5b54725cc76fd568b0058caf7"} Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.316159 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"98d1f3cc0f9ab3d41cdcdd125bb9227106916f474a34592b5e034d6f0c2ba1ac"} Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.316187 4756 scope.go:117] "RemoveContainer" containerID="ef696d62114310b595359fdf0f823a6781a2df097280882ce3528c732940055c" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.335660 4756 generic.go:334] "Generic (PLEG): container finished" podID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerID="a4ed74681d7ba4efc54983a31700306db3a4304eaf38418ef396ea4737d7bc16" exitCode=0 Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.335716 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" event={"ID":"fb9890e6-45ec-47f2-969d-ffed8b40fb5c","Type":"ContainerDied","Data":"a4ed74681d7ba4efc54983a31700306db3a4304eaf38418ef396ea4737d7bc16"} Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.404867 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.456609 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-combined-ca-bundle\") pod \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.456671 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-logs\") pod \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.456730 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data\") pod \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.456758 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data-custom\") pod \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.456789 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j9xp\" (UniqueName: \"kubernetes.io/projected/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-kube-api-access-9j9xp\") pod \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\" (UID: \"fb9890e6-45ec-47f2-969d-ffed8b40fb5c\") " Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.457802 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-logs" (OuterVolumeSpecName: "logs") pod "fb9890e6-45ec-47f2-969d-ffed8b40fb5c" (UID: "fb9890e6-45ec-47f2-969d-ffed8b40fb5c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.463694 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fb9890e6-45ec-47f2-969d-ffed8b40fb5c" (UID: "fb9890e6-45ec-47f2-969d-ffed8b40fb5c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.464712 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-kube-api-access-9j9xp" (OuterVolumeSpecName: "kube-api-access-9j9xp") pod "fb9890e6-45ec-47f2-969d-ffed8b40fb5c" (UID: "fb9890e6-45ec-47f2-969d-ffed8b40fb5c"). InnerVolumeSpecName "kube-api-access-9j9xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.505637 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb9890e6-45ec-47f2-969d-ffed8b40fb5c" (UID: "fb9890e6-45ec-47f2-969d-ffed8b40fb5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.523803 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data" (OuterVolumeSpecName: "config-data") pod "fb9890e6-45ec-47f2-969d-ffed8b40fb5c" (UID: "fb9890e6-45ec-47f2-969d-ffed8b40fb5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.559078 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.559395 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.559407 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.559417 4756 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.559424 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j9xp\" (UniqueName: \"kubernetes.io/projected/fb9890e6-45ec-47f2-969d-ffed8b40fb5c-kube-api-access-9j9xp\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:14 crc kubenswrapper[4756]: I0203 09:29:14.814989 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.350046 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" event={"ID":"fb9890e6-45ec-47f2-969d-ffed8b40fb5c","Type":"ContainerDied","Data":"92e95bc13a65b86da37b2d83c10adf31e5bf7e548939c92189bed3ded00d134a"} Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.350108 4756 scope.go:117] "RemoveContainer" containerID="a4ed74681d7ba4efc54983a31700306db3a4304eaf38418ef396ea4737d7bc16" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.350108 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b7c55c5fb-l7j2x" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.370869 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 03 09:29:15 crc kubenswrapper[4756]: E0203 09:29:15.372035 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.372067 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api" Feb 03 09:29:15 crc kubenswrapper[4756]: E0203 09:29:15.372135 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api-log" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.372146 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api-log" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.372896 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api-log" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.373001 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" containerName="barbican-api" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.374105 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.376825 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.376907 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.380427 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-6zhgb" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.399982 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.415906 4756 scope.go:117] "RemoveContainer" containerID="8fea60c16d0173717be71e1375455613406263a7b257e07d285ac174b84d6127" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.434179 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b7c55c5fb-l7j2x"] Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.441651 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7b7c55c5fb-l7j2x"] Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.478692 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3e4a2949-6bbf-4478-aee7-2cd98095c059-openstack-config-secret\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.478808 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e4a2949-6bbf-4478-aee7-2cd98095c059-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.478841 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3e4a2949-6bbf-4478-aee7-2cd98095c059-openstack-config\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.478901 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8c4w\" (UniqueName: \"kubernetes.io/projected/3e4a2949-6bbf-4478-aee7-2cd98095c059-kube-api-access-t8c4w\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: E0203 09:29:15.572112 4756 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb9890e6_45ec_47f2_969d_ffed8b40fb5c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb9890e6_45ec_47f2_969d_ffed8b40fb5c.slice/crio-92e95bc13a65b86da37b2d83c10adf31e5bf7e548939c92189bed3ded00d134a\": RecentStats: unable to find data in memory cache]" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.580816 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3e4a2949-6bbf-4478-aee7-2cd98095c059-openstack-config-secret\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.580880 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e4a2949-6bbf-4478-aee7-2cd98095c059-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.580911 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3e4a2949-6bbf-4478-aee7-2cd98095c059-openstack-config\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.581069 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8c4w\" (UniqueName: \"kubernetes.io/projected/3e4a2949-6bbf-4478-aee7-2cd98095c059-kube-api-access-t8c4w\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.589265 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3e4a2949-6bbf-4478-aee7-2cd98095c059-openstack-config\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.594525 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3e4a2949-6bbf-4478-aee7-2cd98095c059-openstack-config-secret\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.596219 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e4a2949-6bbf-4478-aee7-2cd98095c059-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.600027 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8c4w\" (UniqueName: \"kubernetes.io/projected/3e4a2949-6bbf-4478-aee7-2cd98095c059-kube-api-access-t8c4w\") pod \"openstackclient\" (UID: \"3e4a2949-6bbf-4478-aee7-2cd98095c059\") " pod="openstack/openstackclient" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.624467 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb9890e6-45ec-47f2-969d-ffed8b40fb5c" path="/var/lib/kubelet/pods/fb9890e6-45ec-47f2-969d-ffed8b40fb5c/volumes" Feb 03 09:29:15 crc kubenswrapper[4756]: I0203 09:29:15.719237 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 03 09:29:16 crc kubenswrapper[4756]: I0203 09:29:16.197367 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 03 09:29:16 crc kubenswrapper[4756]: I0203 09:29:16.359877 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3e4a2949-6bbf-4478-aee7-2cd98095c059","Type":"ContainerStarted","Data":"528f1a510c9b394b4c5b2d360f951af908e5a5ecc1626a1385246436dd268cb2"} Feb 03 09:29:18 crc kubenswrapper[4756]: I0203 09:29:18.165270 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-546db499dd-c5f8j" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.556063 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.556641 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="ceilometer-central-agent" containerID="cri-o://bf36250fb9ab5d06431f72af20bbf4ee9068d16323f2d434636481097391a131" gracePeriod=30 Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.557415 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="proxy-httpd" containerID="cri-o://420fee0f8ce0be289bb2f7da9a5eadb9a5bf8f3a7912485e9a08f7b1a3453a04" gracePeriod=30 Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.557482 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="sg-core" containerID="cri-o://6cd0d88ad9342418a15d0df2f356663f9a991e202daf83f6b323fd429bb10661" gracePeriod=30 Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.557518 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="ceilometer-notification-agent" containerID="cri-o://64793391177c891ddb316dea511f27d059e292e92e021d3f67b7b74a6e7449f5" gracePeriod=30 Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.631331 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-85ff9cc7c5-ns7sz"] Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.632908 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85ff9cc7c5-ns7sz"] Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.632984 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.639511 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.639788 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.639937 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.670009 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.165:3000/\": read tcp 10.217.0.2:54136->10.217.0.165:3000: read: connection reset by peer" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.758551 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74be27bd-98ba-4f2c-bd2e-ea5855812a55-etc-swift\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.758611 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zwdv\" (UniqueName: \"kubernetes.io/projected/74be27bd-98ba-4f2c-bd2e-ea5855812a55-kube-api-access-2zwdv\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.758643 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74be27bd-98ba-4f2c-bd2e-ea5855812a55-log-httpd\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.758685 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-config-data\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.758749 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-internal-tls-certs\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.758770 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74be27bd-98ba-4f2c-bd2e-ea5855812a55-run-httpd\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.758818 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-public-tls-certs\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.758836 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-combined-ca-bundle\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.860491 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-config-data\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.860581 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-internal-tls-certs\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.860611 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74be27bd-98ba-4f2c-bd2e-ea5855812a55-run-httpd\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.860668 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-public-tls-certs\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.860685 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-combined-ca-bundle\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.860722 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74be27bd-98ba-4f2c-bd2e-ea5855812a55-etc-swift\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.860742 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zwdv\" (UniqueName: \"kubernetes.io/projected/74be27bd-98ba-4f2c-bd2e-ea5855812a55-kube-api-access-2zwdv\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.860773 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74be27bd-98ba-4f2c-bd2e-ea5855812a55-log-httpd\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.861224 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74be27bd-98ba-4f2c-bd2e-ea5855812a55-log-httpd\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.862064 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74be27bd-98ba-4f2c-bd2e-ea5855812a55-run-httpd\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.867542 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-public-tls-certs\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.868381 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/74be27bd-98ba-4f2c-bd2e-ea5855812a55-etc-swift\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.870302 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-internal-tls-certs\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.872639 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-combined-ca-bundle\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.877491 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zwdv\" (UniqueName: \"kubernetes.io/projected/74be27bd-98ba-4f2c-bd2e-ea5855812a55-kube-api-access-2zwdv\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.886234 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74be27bd-98ba-4f2c-bd2e-ea5855812a55-config-data\") pod \"swift-proxy-85ff9cc7c5-ns7sz\" (UID: \"74be27bd-98ba-4f2c-bd2e-ea5855812a55\") " pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:19 crc kubenswrapper[4756]: I0203 09:29:19.963867 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:20 crc kubenswrapper[4756]: I0203 09:29:20.134264 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 03 09:29:20 crc kubenswrapper[4756]: I0203 09:29:20.396774 4756 generic.go:334] "Generic (PLEG): container finished" podID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerID="420fee0f8ce0be289bb2f7da9a5eadb9a5bf8f3a7912485e9a08f7b1a3453a04" exitCode=0 Feb 03 09:29:20 crc kubenswrapper[4756]: I0203 09:29:20.396803 4756 generic.go:334] "Generic (PLEG): container finished" podID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerID="6cd0d88ad9342418a15d0df2f356663f9a991e202daf83f6b323fd429bb10661" exitCode=2 Feb 03 09:29:20 crc kubenswrapper[4756]: I0203 09:29:20.396812 4756 generic.go:334] "Generic (PLEG): container finished" podID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerID="bf36250fb9ab5d06431f72af20bbf4ee9068d16323f2d434636481097391a131" exitCode=0 Feb 03 09:29:20 crc kubenswrapper[4756]: I0203 09:29:20.396830 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerDied","Data":"420fee0f8ce0be289bb2f7da9a5eadb9a5bf8f3a7912485e9a08f7b1a3453a04"} Feb 03 09:29:20 crc kubenswrapper[4756]: I0203 09:29:20.396855 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerDied","Data":"6cd0d88ad9342418a15d0df2f356663f9a991e202daf83f6b323fd429bb10661"} Feb 03 09:29:20 crc kubenswrapper[4756]: I0203 09:29:20.396865 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerDied","Data":"bf36250fb9ab5d06431f72af20bbf4ee9068d16323f2d434636481097391a131"} Feb 03 09:29:20 crc kubenswrapper[4756]: I0203 09:29:20.561288 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-85ff9cc7c5-ns7sz"] Feb 03 09:29:23 crc kubenswrapper[4756]: I0203 09:29:23.426482 4756 generic.go:334] "Generic (PLEG): container finished" podID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerID="64793391177c891ddb316dea511f27d059e292e92e021d3f67b7b74a6e7449f5" exitCode=0 Feb 03 09:29:23 crc kubenswrapper[4756]: I0203 09:29:23.426573 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerDied","Data":"64793391177c891ddb316dea511f27d059e292e92e021d3f67b7b74a6e7449f5"} Feb 03 09:29:23 crc kubenswrapper[4756]: I0203 09:29:23.906471 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.165:3000/\": dial tcp 10.217.0.165:3000: connect: connection refused" Feb 03 09:29:25 crc kubenswrapper[4756]: W0203 09:29:25.008762 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74be27bd_98ba_4f2c_bd2e_ea5855812a55.slice/crio-4d14ff2fdbd3b5c99d5fe320aceaad8ae9aeacfb0683744c9fa3d0f0938a2f61 WatchSource:0}: Error finding container 4d14ff2fdbd3b5c99d5fe320aceaad8ae9aeacfb0683744c9fa3d0f0938a2f61: Status 404 returned error can't find the container with id 4d14ff2fdbd3b5c99d5fe320aceaad8ae9aeacfb0683744c9fa3d0f0938a2f61 Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.323544 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.365652 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-scripts\") pod \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.365717 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-sg-core-conf-yaml\") pod \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.365793 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-run-httpd\") pod \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.365908 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrcd8\" (UniqueName: \"kubernetes.io/projected/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-kube-api-access-mrcd8\") pod \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.365934 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-config-data\") pod \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.365980 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-log-httpd\") pod \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.366151 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-combined-ca-bundle\") pod \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\" (UID: \"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5\") " Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.366468 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" (UID: "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.366513 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" (UID: "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.367049 4756 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.367078 4756 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.372084 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-scripts" (OuterVolumeSpecName: "scripts") pod "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" (UID: "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.375907 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-kube-api-access-mrcd8" (OuterVolumeSpecName: "kube-api-access-mrcd8") pod "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" (UID: "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5"). InnerVolumeSpecName "kube-api-access-mrcd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.419572 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" (UID: "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.470311 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.471599 4756 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.471695 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrcd8\" (UniqueName: \"kubernetes.io/projected/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-kube-api-access-mrcd8\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.479410 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.479407 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5","Type":"ContainerDied","Data":"a803c8442edc19e01a55b0ab1fa906e01a6acc8073cb737d069588c1698c1482"} Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.479808 4756 scope.go:117] "RemoveContainer" containerID="420fee0f8ce0be289bb2f7da9a5eadb9a5bf8f3a7912485e9a08f7b1a3453a04" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.488650 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.488997 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-log" containerID="cri-o://910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0" gracePeriod=30 Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.489565 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-httpd" containerID="cri-o://86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30" gracePeriod=30 Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.494867 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" event={"ID":"74be27bd-98ba-4f2c-bd2e-ea5855812a55","Type":"ContainerStarted","Data":"dbd999b2835a775c8758c24b71106d8532af900a934ab7f7c22790a497b6d608"} Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.495784 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" event={"ID":"74be27bd-98ba-4f2c-bd2e-ea5855812a55","Type":"ContainerStarted","Data":"4d14ff2fdbd3b5c99d5fe320aceaad8ae9aeacfb0683744c9fa3d0f0938a2f61"} Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.500824 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" (UID: "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.501362 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3e4a2949-6bbf-4478-aee7-2cd98095c059","Type":"ContainerStarted","Data":"94237f8f75376c63088fec1d4bfee1f07caa2288676c8b0133cfe82d5a5c19e9"} Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.525925 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.6123911720000002 podStartE2EDuration="10.525903844s" podCreationTimestamp="2026-02-03 09:29:15 +0000 UTC" firstStartedPulling="2026-02-03 09:29:16.210010216 +0000 UTC m=+1147.360477591" lastFinishedPulling="2026-02-03 09:29:25.123522888 +0000 UTC m=+1156.273990263" observedRunningTime="2026-02-03 09:29:25.521078792 +0000 UTC m=+1156.671546167" watchObservedRunningTime="2026-02-03 09:29:25.525903844 +0000 UTC m=+1156.676371219" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.542770 4756 scope.go:117] "RemoveContainer" containerID="6cd0d88ad9342418a15d0df2f356663f9a991e202daf83f6b323fd429bb10661" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.565255 4756 scope.go:117] "RemoveContainer" containerID="64793391177c891ddb316dea511f27d059e292e92e021d3f67b7b74a6e7449f5" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.573369 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.584712 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-config-data" (OuterVolumeSpecName: "config-data") pod "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" (UID: "ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.589527 4756 scope.go:117] "RemoveContainer" containerID="bf36250fb9ab5d06431f72af20bbf4ee9068d16323f2d434636481097391a131" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.675014 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.816769 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.836569 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.862435 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:25 crc kubenswrapper[4756]: E0203 09:29:25.862860 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="ceilometer-notification-agent" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.862876 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="ceilometer-notification-agent" Feb 03 09:29:25 crc kubenswrapper[4756]: E0203 09:29:25.862888 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="proxy-httpd" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.862895 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="proxy-httpd" Feb 03 09:29:25 crc kubenswrapper[4756]: E0203 09:29:25.862906 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="ceilometer-central-agent" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.862911 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="ceilometer-central-agent" Feb 03 09:29:25 crc kubenswrapper[4756]: E0203 09:29:25.862924 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="sg-core" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.862930 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="sg-core" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.863079 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="ceilometer-central-agent" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.863093 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="sg-core" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.863104 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="proxy-httpd" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.863111 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" containerName="ceilometer-notification-agent" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.864600 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.866840 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.867767 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.876726 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.984102 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.984166 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-scripts\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.984282 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-config-data\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.984312 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-log-httpd\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.984393 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.984419 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-run-httpd\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:25 crc kubenswrapper[4756]: I0203 09:29:25.984470 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhvpx\" (UniqueName: \"kubernetes.io/projected/84b66417-49cf-498b-ab27-3a8872227641-kube-api-access-fhvpx\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.040110 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-x676l"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.046921 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.053998 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x676l"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.089622 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.089901 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89hrk\" (UniqueName: \"kubernetes.io/projected/95eac2ad-73dc-405d-b583-92d263bec833-kube-api-access-89hrk\") pod \"nova-api-db-create-x676l\" (UID: \"95eac2ad-73dc-405d-b583-92d263bec833\") " pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.090018 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-scripts\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.090146 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95eac2ad-73dc-405d-b583-92d263bec833-operator-scripts\") pod \"nova-api-db-create-x676l\" (UID: \"95eac2ad-73dc-405d-b583-92d263bec833\") " pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.090282 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-config-data\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.090675 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-log-httpd\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.090859 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.090962 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-run-httpd\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.091072 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhvpx\" (UniqueName: \"kubernetes.io/projected/84b66417-49cf-498b-ab27-3a8872227641-kube-api-access-fhvpx\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.091221 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-log-httpd\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.091854 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-run-httpd\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.095814 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-scripts\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.096556 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-config-data\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.097207 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.097535 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.110129 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhvpx\" (UniqueName: \"kubernetes.io/projected/84b66417-49cf-498b-ab27-3a8872227641-kube-api-access-fhvpx\") pod \"ceilometer-0\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.143844 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-6s59j"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.145323 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.163287 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6s59j"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.175501 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5fd1-account-create-update-6kqzn"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.176844 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.179130 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.184671 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.185057 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5fd1-account-create-update-6kqzn"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.192732 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89hrk\" (UniqueName: \"kubernetes.io/projected/95eac2ad-73dc-405d-b583-92d263bec833-kube-api-access-89hrk\") pod \"nova-api-db-create-x676l\" (UID: \"95eac2ad-73dc-405d-b583-92d263bec833\") " pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.192826 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95eac2ad-73dc-405d-b583-92d263bec833-operator-scripts\") pod \"nova-api-db-create-x676l\" (UID: \"95eac2ad-73dc-405d-b583-92d263bec833\") " pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.192935 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24066fdc-6998-4f09-9a07-230cb3378f11-operator-scripts\") pod \"nova-cell0-db-create-6s59j\" (UID: \"24066fdc-6998-4f09-9a07-230cb3378f11\") " pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.192988 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rbvb\" (UniqueName: \"kubernetes.io/projected/24066fdc-6998-4f09-9a07-230cb3378f11-kube-api-access-6rbvb\") pod \"nova-cell0-db-create-6s59j\" (UID: \"24066fdc-6998-4f09-9a07-230cb3378f11\") " pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.193792 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95eac2ad-73dc-405d-b583-92d263bec833-operator-scripts\") pod \"nova-api-db-create-x676l\" (UID: \"95eac2ad-73dc-405d-b583-92d263bec833\") " pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.219242 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89hrk\" (UniqueName: \"kubernetes.io/projected/95eac2ad-73dc-405d-b583-92d263bec833-kube-api-access-89hrk\") pod \"nova-api-db-create-x676l\" (UID: \"95eac2ad-73dc-405d-b583-92d263bec833\") " pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.257472 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.257737 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-log" containerID="cri-o://0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef" gracePeriod=30 Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.258242 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-httpd" containerID="cri-o://5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b" gracePeriod=30 Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.295725 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdpwn\" (UniqueName: \"kubernetes.io/projected/4172aecf-8ba4-4990-9bd9-11ef49b0320e-kube-api-access-bdpwn\") pod \"nova-api-5fd1-account-create-update-6kqzn\" (UID: \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\") " pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.295813 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4172aecf-8ba4-4990-9bd9-11ef49b0320e-operator-scripts\") pod \"nova-api-5fd1-account-create-update-6kqzn\" (UID: \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\") " pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.295848 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24066fdc-6998-4f09-9a07-230cb3378f11-operator-scripts\") pod \"nova-cell0-db-create-6s59j\" (UID: \"24066fdc-6998-4f09-9a07-230cb3378f11\") " pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.295884 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rbvb\" (UniqueName: \"kubernetes.io/projected/24066fdc-6998-4f09-9a07-230cb3378f11-kube-api-access-6rbvb\") pod \"nova-cell0-db-create-6s59j\" (UID: \"24066fdc-6998-4f09-9a07-230cb3378f11\") " pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.308628 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.309968 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24066fdc-6998-4f09-9a07-230cb3378f11-operator-scripts\") pod \"nova-cell0-db-create-6s59j\" (UID: \"24066fdc-6998-4f09-9a07-230cb3378f11\") " pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.321797 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rbvb\" (UniqueName: \"kubernetes.io/projected/24066fdc-6998-4f09-9a07-230cb3378f11-kube-api-access-6rbvb\") pod \"nova-cell0-db-create-6s59j\" (UID: \"24066fdc-6998-4f09-9a07-230cb3378f11\") " pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.337181 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.384698 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f912-account-create-update-t2sfb"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.386590 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.390918 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.397202 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdpwn\" (UniqueName: \"kubernetes.io/projected/4172aecf-8ba4-4990-9bd9-11ef49b0320e-kube-api-access-bdpwn\") pod \"nova-api-5fd1-account-create-update-6kqzn\" (UID: \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\") " pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.397292 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4172aecf-8ba4-4990-9bd9-11ef49b0320e-operator-scripts\") pod \"nova-api-5fd1-account-create-update-6kqzn\" (UID: \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\") " pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.399220 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4172aecf-8ba4-4990-9bd9-11ef49b0320e-operator-scripts\") pod \"nova-api-5fd1-account-create-update-6kqzn\" (UID: \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\") " pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.399722 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-d2m25"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.400764 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.424495 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f912-account-create-update-t2sfb"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.431950 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdpwn\" (UniqueName: \"kubernetes.io/projected/4172aecf-8ba4-4990-9bd9-11ef49b0320e-kube-api-access-bdpwn\") pod \"nova-api-5fd1-account-create-update-6kqzn\" (UID: \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\") " pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.432026 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-d2m25"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.499475 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm486\" (UniqueName: \"kubernetes.io/projected/15a12c2e-5874-45c7-bc27-1281125fb024-kube-api-access-rm486\") pod \"nova-cell1-db-create-d2m25\" (UID: \"15a12c2e-5874-45c7-bc27-1281125fb024\") " pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.499556 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgt5k\" (UniqueName: \"kubernetes.io/projected/a2b922d8-73c3-42e7-a7af-f28120f67125-kube-api-access-lgt5k\") pod \"nova-cell0-f912-account-create-update-t2sfb\" (UID: \"a2b922d8-73c3-42e7-a7af-f28120f67125\") " pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.499654 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a12c2e-5874-45c7-bc27-1281125fb024-operator-scripts\") pod \"nova-cell1-db-create-d2m25\" (UID: \"15a12c2e-5874-45c7-bc27-1281125fb024\") " pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.499761 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b922d8-73c3-42e7-a7af-f28120f67125-operator-scripts\") pod \"nova-cell0-f912-account-create-update-t2sfb\" (UID: \"a2b922d8-73c3-42e7-a7af-f28120f67125\") " pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.561434 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" event={"ID":"74be27bd-98ba-4f2c-bd2e-ea5855812a55","Type":"ContainerStarted","Data":"dcc7e4802d9abb7fe48a0e0c8eacec9c9f4c70cdf375f9af82297cc7a5a44691"} Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.563023 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.563103 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.577970 4756 generic.go:334] "Generic (PLEG): container finished" podID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerID="910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0" exitCode=143 Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.578072 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"19ccb5fd-4e2d-4cc8-b776-161687b99eaa","Type":"ContainerDied","Data":"910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0"} Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.592288 4756 generic.go:334] "Generic (PLEG): container finished" podID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerID="0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef" exitCode=143 Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.592487 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16c4732e-295a-495d-a2af-e10b2c71a10f","Type":"ContainerDied","Data":"0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef"} Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.595962 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f9ed-account-create-update-jnkxm"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.598500 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.600367 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.602703 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a12c2e-5874-45c7-bc27-1281125fb024-operator-scripts\") pod \"nova-cell1-db-create-d2m25\" (UID: \"15a12c2e-5874-45c7-bc27-1281125fb024\") " pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.602887 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b922d8-73c3-42e7-a7af-f28120f67125-operator-scripts\") pod \"nova-cell0-f912-account-create-update-t2sfb\" (UID: \"a2b922d8-73c3-42e7-a7af-f28120f67125\") " pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.602976 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm486\" (UniqueName: \"kubernetes.io/projected/15a12c2e-5874-45c7-bc27-1281125fb024-kube-api-access-rm486\") pod \"nova-cell1-db-create-d2m25\" (UID: \"15a12c2e-5874-45c7-bc27-1281125fb024\") " pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.603050 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgt5k\" (UniqueName: \"kubernetes.io/projected/a2b922d8-73c3-42e7-a7af-f28120f67125-kube-api-access-lgt5k\") pod \"nova-cell0-f912-account-create-update-t2sfb\" (UID: \"a2b922d8-73c3-42e7-a7af-f28120f67125\") " pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.603535 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b922d8-73c3-42e7-a7af-f28120f67125-operator-scripts\") pod \"nova-cell0-f912-account-create-update-t2sfb\" (UID: \"a2b922d8-73c3-42e7-a7af-f28120f67125\") " pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.603828 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a12c2e-5874-45c7-bc27-1281125fb024-operator-scripts\") pod \"nova-cell1-db-create-d2m25\" (UID: \"15a12c2e-5874-45c7-bc27-1281125fb024\") " pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.605098 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f9ed-account-create-update-jnkxm"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.606098 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" podStartSLOduration=7.606082969 podStartE2EDuration="7.606082969s" podCreationTimestamp="2026-02-03 09:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:26.588307793 +0000 UTC m=+1157.738775158" watchObservedRunningTime="2026-02-03 09:29:26.606082969 +0000 UTC m=+1157.756550344" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.624922 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgt5k\" (UniqueName: \"kubernetes.io/projected/a2b922d8-73c3-42e7-a7af-f28120f67125-kube-api-access-lgt5k\") pod \"nova-cell0-f912-account-create-update-t2sfb\" (UID: \"a2b922d8-73c3-42e7-a7af-f28120f67125\") " pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.625843 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm486\" (UniqueName: \"kubernetes.io/projected/15a12c2e-5874-45c7-bc27-1281125fb024-kube-api-access-rm486\") pod \"nova-cell1-db-create-d2m25\" (UID: \"15a12c2e-5874-45c7-bc27-1281125fb024\") " pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.681501 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.711024 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7dgf\" (UniqueName: \"kubernetes.io/projected/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-kube-api-access-x7dgf\") pod \"nova-cell1-f9ed-account-create-update-jnkxm\" (UID: \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\") " pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.711589 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-operator-scripts\") pod \"nova-cell1-f9ed-account-create-update-jnkxm\" (UID: \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\") " pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.729735 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.768705 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.817253 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-operator-scripts\") pod \"nova-cell1-f9ed-account-create-update-jnkxm\" (UID: \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\") " pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.817672 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7dgf\" (UniqueName: \"kubernetes.io/projected/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-kube-api-access-x7dgf\") pod \"nova-cell1-f9ed-account-create-update-jnkxm\" (UID: \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\") " pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.818431 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-operator-scripts\") pod \"nova-cell1-f9ed-account-create-update-jnkxm\" (UID: \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\") " pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.832579 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.857943 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7dgf\" (UniqueName: \"kubernetes.io/projected/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-kube-api-access-x7dgf\") pod \"nova-cell1-f9ed-account-create-update-jnkxm\" (UID: \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\") " pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.948929 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:26 crc kubenswrapper[4756]: I0203 09:29:26.994938 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-x676l"] Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.006041 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6s59j"] Feb 03 09:29:27 crc kubenswrapper[4756]: W0203 09:29:27.022973 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24066fdc_6998_4f09_9a07_230cb3378f11.slice/crio-c57ab2175efbd77ee462bd3480ce8f25adb9cec38e7cf75dd0aadb3770fceb60 WatchSource:0}: Error finding container c57ab2175efbd77ee462bd3480ce8f25adb9cec38e7cf75dd0aadb3770fceb60: Status 404 returned error can't find the container with id c57ab2175efbd77ee462bd3480ce8f25adb9cec38e7cf75dd0aadb3770fceb60 Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.266518 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5fd1-account-create-update-6kqzn"] Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.285181 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f912-account-create-update-t2sfb"] Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.563327 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-d2m25"] Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.635382 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5" path="/var/lib/kubelet/pods/ee2fadc1-7b6c-4d19-abaf-57ece18ec2c5/volumes" Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.644257 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerStarted","Data":"64fd379a227ea71f5d7f14bfffe85f5c40be5977f76e093b024338b1c7c447db"} Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.655558 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d2m25" event={"ID":"15a12c2e-5874-45c7-bc27-1281125fb024","Type":"ContainerStarted","Data":"e14b8f867734a8d4fb06a3744eb7e3c6b016cc0714f0b81789b977e160d20ce4"} Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.669465 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5fd1-account-create-update-6kqzn" event={"ID":"4172aecf-8ba4-4990-9bd9-11ef49b0320e","Type":"ContainerStarted","Data":"0d906a731ab533d2aa77d84e5140bd916f0dbeb31f9e0ab29e7bc94e1bf134cb"} Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.691934 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x676l" event={"ID":"95eac2ad-73dc-405d-b583-92d263bec833","Type":"ContainerStarted","Data":"a654d32d14744b248ff72832f42e1ca11792dd1ab1ab25b9372dcfafe604a53d"} Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.691986 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x676l" event={"ID":"95eac2ad-73dc-405d-b583-92d263bec833","Type":"ContainerStarted","Data":"0d458f75237d04c9dc0d95e438455e4cf81ccc5faf301ea1f1715d7ac55b2794"} Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.697572 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f912-account-create-update-t2sfb" event={"ID":"a2b922d8-73c3-42e7-a7af-f28120f67125","Type":"ContainerStarted","Data":"6d5696e7e9b8f6bd6f7370707e315eb4d2210eae0ea148376b7d4a4eaa00ef57"} Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.704227 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6s59j" event={"ID":"24066fdc-6998-4f09-9a07-230cb3378f11","Type":"ContainerStarted","Data":"d2f60bbc9e7f4bbd60b771f014093defccf5e361ff795b00ee4e450d490222d0"} Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.704268 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6s59j" event={"ID":"24066fdc-6998-4f09-9a07-230cb3378f11","Type":"ContainerStarted","Data":"c57ab2175efbd77ee462bd3480ce8f25adb9cec38e7cf75dd0aadb3770fceb60"} Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.729506 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f9ed-account-create-update-jnkxm"] Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.739969 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-x676l" podStartSLOduration=1.7399483629999999 podStartE2EDuration="1.739948363s" podCreationTimestamp="2026-02-03 09:29:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:27.715028034 +0000 UTC m=+1158.865495409" watchObservedRunningTime="2026-02-03 09:29:27.739948363 +0000 UTC m=+1158.890415748" Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.756005 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-6s59j" podStartSLOduration=1.7559896350000002 podStartE2EDuration="1.755989635s" podCreationTimestamp="2026-02-03 09:29:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:27.734791042 +0000 UTC m=+1158.885258417" watchObservedRunningTime="2026-02-03 09:29:27.755989635 +0000 UTC m=+1158.906457010" Feb 03 09:29:27 crc kubenswrapper[4756]: I0203 09:29:27.988892 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8547fc4b5c-w4sjb" Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.085807 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7b74899c58-bsxwf"] Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.087634 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7b74899c58-bsxwf" podUID="c571905e-5655-45c7-a864-70d02e7a9326" containerName="neutron-httpd" containerID="cri-o://dedf5fd2bb1e43cfe9147a8997060b222a345753833e11dab8117b21d7572e1d" gracePeriod=30 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.086097 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7b74899c58-bsxwf" podUID="c571905e-5655-45c7-a864-70d02e7a9326" containerName="neutron-api" containerID="cri-o://1b0973a2612cd6f68029aef4d5a1476dd1e24253e75900f5476116c0a5b7eca0" gracePeriod=30 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.168691 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-546db499dd-c5f8j" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.547272 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.703028 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.151:9292/healthcheck\": read tcp 10.217.0.2:54188->10.217.0.151:9292: read: connection reset by peer" Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.703027 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.151:9292/healthcheck\": read tcp 10.217.0.2:54202->10.217.0.151:9292: read: connection reset by peer" Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.716630 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerStarted","Data":"f23a5200fcfb1ed2593606e2a184dd0f9adaade347b226fcad6da01d47f764b4"} Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.719457 4756 generic.go:334] "Generic (PLEG): container finished" podID="c571905e-5655-45c7-a864-70d02e7a9326" containerID="dedf5fd2bb1e43cfe9147a8997060b222a345753833e11dab8117b21d7572e1d" exitCode=0 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.719533 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b74899c58-bsxwf" event={"ID":"c571905e-5655-45c7-a864-70d02e7a9326","Type":"ContainerDied","Data":"dedf5fd2bb1e43cfe9147a8997060b222a345753833e11dab8117b21d7572e1d"} Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.722014 4756 generic.go:334] "Generic (PLEG): container finished" podID="95eac2ad-73dc-405d-b583-92d263bec833" containerID="a654d32d14744b248ff72832f42e1ca11792dd1ab1ab25b9372dcfafe604a53d" exitCode=0 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.722064 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x676l" event={"ID":"95eac2ad-73dc-405d-b583-92d263bec833","Type":"ContainerDied","Data":"a654d32d14744b248ff72832f42e1ca11792dd1ab1ab25b9372dcfafe604a53d"} Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.726082 4756 generic.go:334] "Generic (PLEG): container finished" podID="4172aecf-8ba4-4990-9bd9-11ef49b0320e" containerID="35ceb7ea9786f3b105817c46c6f20f572d4769869bb13a1ce23f96fc639c759c" exitCode=0 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.726134 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5fd1-account-create-update-6kqzn" event={"ID":"4172aecf-8ba4-4990-9bd9-11ef49b0320e","Type":"ContainerDied","Data":"35ceb7ea9786f3b105817c46c6f20f572d4769869bb13a1ce23f96fc639c759c"} Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.730758 4756 generic.go:334] "Generic (PLEG): container finished" podID="15a12c2e-5874-45c7-bc27-1281125fb024" containerID="aad8c853f8c178c99a92838cbe7181f9283f11434ba6a9a3593f17c77d43dabe" exitCode=0 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.730782 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d2m25" event={"ID":"15a12c2e-5874-45c7-bc27-1281125fb024","Type":"ContainerDied","Data":"aad8c853f8c178c99a92838cbe7181f9283f11434ba6a9a3593f17c77d43dabe"} Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.734118 4756 generic.go:334] "Generic (PLEG): container finished" podID="d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea" containerID="67abb070992b5549ffeba6a615e778eb71f9e9e5e9cf9624ab62546df59517b2" exitCode=0 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.734153 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" event={"ID":"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea","Type":"ContainerDied","Data":"67abb070992b5549ffeba6a615e778eb71f9e9e5e9cf9624ab62546df59517b2"} Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.734167 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" event={"ID":"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea","Type":"ContainerStarted","Data":"1d83dd66fedd7334842056b52516c8e7a755fda81e9de4a297c99ee357496050"} Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.742039 4756 generic.go:334] "Generic (PLEG): container finished" podID="a2b922d8-73c3-42e7-a7af-f28120f67125" containerID="2fdbf7fc7b72893e059e77d00d4c50b542d97835f099d9f579047043b519cb79" exitCode=0 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.742127 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f912-account-create-update-t2sfb" event={"ID":"a2b922d8-73c3-42e7-a7af-f28120f67125","Type":"ContainerDied","Data":"2fdbf7fc7b72893e059e77d00d4c50b542d97835f099d9f579047043b519cb79"} Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.749863 4756 generic.go:334] "Generic (PLEG): container finished" podID="24066fdc-6998-4f09-9a07-230cb3378f11" containerID="d2f60bbc9e7f4bbd60b771f014093defccf5e361ff795b00ee4e450d490222d0" exitCode=0 Feb 03 09:29:28 crc kubenswrapper[4756]: I0203 09:29:28.749923 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6s59j" event={"ID":"24066fdc-6998-4f09-9a07-230cb3378f11","Type":"ContainerDied","Data":"d2f60bbc9e7f4bbd60b771f014093defccf5e361ff795b00ee4e450d490222d0"} Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.274198 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.352296 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.409316 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-etc-machine-id\") pod \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.409453 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fd03b9b6-ff6d-4eed-be6d-e08461b6c693" (UID: "fd03b9b6-ff6d-4eed-be6d-e08461b6c693"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.409532 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data-custom\") pod \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.409621 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpj9w\" (UniqueName: \"kubernetes.io/projected/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-kube-api-access-tpj9w\") pod \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.409669 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-logs\") pod \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.409723 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data\") pod \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.409739 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-scripts\") pod \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.409785 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-combined-ca-bundle\") pod \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\" (UID: \"fd03b9b6-ff6d-4eed-be6d-e08461b6c693\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.410211 4756 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.410315 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-logs" (OuterVolumeSpecName: "logs") pod "fd03b9b6-ff6d-4eed-be6d-e08461b6c693" (UID: "fd03b9b6-ff6d-4eed-be6d-e08461b6c693"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.414335 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fd03b9b6-ff6d-4eed-be6d-e08461b6c693" (UID: "fd03b9b6-ff6d-4eed-be6d-e08461b6c693"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.417598 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-kube-api-access-tpj9w" (OuterVolumeSpecName: "kube-api-access-tpj9w") pod "fd03b9b6-ff6d-4eed-be6d-e08461b6c693" (UID: "fd03b9b6-ff6d-4eed-be6d-e08461b6c693"). InnerVolumeSpecName "kube-api-access-tpj9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.420551 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-scripts" (OuterVolumeSpecName: "scripts") pod "fd03b9b6-ff6d-4eed-be6d-e08461b6c693" (UID: "fd03b9b6-ff6d-4eed-be6d-e08461b6c693"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.456152 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd03b9b6-ff6d-4eed-be6d-e08461b6c693" (UID: "fd03b9b6-ff6d-4eed-be6d-e08461b6c693"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.481145 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data" (OuterVolumeSpecName: "config-data") pod "fd03b9b6-ff6d-4eed-be6d-e08461b6c693" (UID: "fd03b9b6-ff6d-4eed-be6d-e08461b6c693"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.511673 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-httpd-run\") pod \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.511869 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-config-data\") pod \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.511919 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8tv6\" (UniqueName: \"kubernetes.io/projected/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-kube-api-access-l8tv6\") pod \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.511946 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-public-tls-certs\") pod \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512011 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-scripts\") pod \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512069 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-logs\") pod \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512142 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512179 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-combined-ca-bundle\") pod \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\" (UID: \"19ccb5fd-4e2d-4cc8-b776-161687b99eaa\") " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512774 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpj9w\" (UniqueName: \"kubernetes.io/projected/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-kube-api-access-tpj9w\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512794 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512806 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512816 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512826 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.512840 4756 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd03b9b6-ff6d-4eed-be6d-e08461b6c693-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.517225 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-scripts" (OuterVolumeSpecName: "scripts") pod "19ccb5fd-4e2d-4cc8-b776-161687b99eaa" (UID: "19ccb5fd-4e2d-4cc8-b776-161687b99eaa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.517548 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-logs" (OuterVolumeSpecName: "logs") pod "19ccb5fd-4e2d-4cc8-b776-161687b99eaa" (UID: "19ccb5fd-4e2d-4cc8-b776-161687b99eaa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.518178 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "19ccb5fd-4e2d-4cc8-b776-161687b99eaa" (UID: "19ccb5fd-4e2d-4cc8-b776-161687b99eaa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.522031 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-kube-api-access-l8tv6" (OuterVolumeSpecName: "kube-api-access-l8tv6") pod "19ccb5fd-4e2d-4cc8-b776-161687b99eaa" (UID: "19ccb5fd-4e2d-4cc8-b776-161687b99eaa"). InnerVolumeSpecName "kube-api-access-l8tv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.527598 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "19ccb5fd-4e2d-4cc8-b776-161687b99eaa" (UID: "19ccb5fd-4e2d-4cc8-b776-161687b99eaa"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.586774 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19ccb5fd-4e2d-4cc8-b776-161687b99eaa" (UID: "19ccb5fd-4e2d-4cc8-b776-161687b99eaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.589031 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "19ccb5fd-4e2d-4cc8-b776-161687b99eaa" (UID: "19ccb5fd-4e2d-4cc8-b776-161687b99eaa"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.600162 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-config-data" (OuterVolumeSpecName: "config-data") pod "19ccb5fd-4e2d-4cc8-b776-161687b99eaa" (UID: "19ccb5fd-4e2d-4cc8-b776-161687b99eaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.616146 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.616177 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.616210 4756 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.616225 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.616239 4756 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.616249 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.616262 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8tv6\" (UniqueName: \"kubernetes.io/projected/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-kube-api-access-l8tv6\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.616273 4756 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/19ccb5fd-4e2d-4cc8-b776-161687b99eaa-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.652138 4756 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.697686 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:48786->10.217.0.152:9292: read: connection reset by peer" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.698007 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:48778->10.217.0.152:9292: read: connection reset by peer" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.717767 4756 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.766892 4756 generic.go:334] "Generic (PLEG): container finished" podID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerID="34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2" exitCode=137 Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.766957 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd03b9b6-ff6d-4eed-be6d-e08461b6c693","Type":"ContainerDied","Data":"34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2"} Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.766982 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fd03b9b6-ff6d-4eed-be6d-e08461b6c693","Type":"ContainerDied","Data":"ef23578273bcfc98a4a44983f488407a2eaef76a5b1b615cb72f99ddab3850d9"} Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.766998 4756 scope.go:117] "RemoveContainer" containerID="34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.767098 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.784161 4756 generic.go:334] "Generic (PLEG): container finished" podID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerID="86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30" exitCode=0 Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.784272 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.784296 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"19ccb5fd-4e2d-4cc8-b776-161687b99eaa","Type":"ContainerDied","Data":"86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30"} Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.785311 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"19ccb5fd-4e2d-4cc8-b776-161687b99eaa","Type":"ContainerDied","Data":"137b54407711507c81c2f21c0e61b626e51715a3e585a15b9ba37a5327a3fe98"} Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.788948 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerStarted","Data":"2e54bad7a03e5dc604442aba7be16afeb3ed24cbd85f34680f4cce5c88096e95"} Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.824498 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.889738 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.930833 4756 scope.go:117] "RemoveContainer" containerID="f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.930948 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.941519 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.962463 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:29:29 crc kubenswrapper[4756]: E0203 09:29:29.962787 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerName="cinder-api-log" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.962798 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerName="cinder-api-log" Feb 03 09:29:29 crc kubenswrapper[4756]: E0203 09:29:29.962814 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-log" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.962820 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-log" Feb 03 09:29:29 crc kubenswrapper[4756]: E0203 09:29:29.962835 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-httpd" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.962841 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-httpd" Feb 03 09:29:29 crc kubenswrapper[4756]: E0203 09:29:29.962850 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerName="cinder-api" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.962856 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerName="cinder-api" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.963020 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerName="cinder-api" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.963033 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" containerName="cinder-api-log" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.963050 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-log" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.963062 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" containerName="glance-httpd" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.963958 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.966965 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.967149 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.967771 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.979881 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.981611 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.984239 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 03 09:29:29 crc kubenswrapper[4756]: I0203 09:29:29.984416 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.040971 4756 scope.go:117] "RemoveContainer" containerID="34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2" Feb 03 09:29:30 crc kubenswrapper[4756]: E0203 09:29:30.041700 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2\": container with ID starting with 34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2 not found: ID does not exist" containerID="34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.041754 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2"} err="failed to get container status \"34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2\": rpc error: code = NotFound desc = could not find container \"34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2\": container with ID starting with 34cc362fc51030fdeca5d2dad2682d5b95980f04a7cf4bb30efd49a94df54fb2 not found: ID does not exist" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.041783 4756 scope.go:117] "RemoveContainer" containerID="f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93" Feb 03 09:29:30 crc kubenswrapper[4756]: E0203 09:29:30.042393 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93\": container with ID starting with f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93 not found: ID does not exist" containerID="f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.042439 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93"} err="failed to get container status \"f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93\": rpc error: code = NotFound desc = could not find container \"f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93\": container with ID starting with f41532e58035efec645c41012414d9afddfd1db057ec9f7ecd703511ee019d93 not found: ID does not exist" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.053867 4756 scope.go:117] "RemoveContainer" containerID="86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.059005 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134151 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg4vc\" (UniqueName: \"kubernetes.io/projected/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-kube-api-access-xg4vc\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134223 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-scripts\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134266 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s9p9\" (UniqueName: \"kubernetes.io/projected/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-kube-api-access-2s9p9\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134292 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134327 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-config-data\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134350 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134377 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-logs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134409 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134434 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134512 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134537 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-config-data\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134562 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134594 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-scripts\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.134620 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-logs\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.136229 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.137574 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.137643 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.137675 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239173 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg4vc\" (UniqueName: \"kubernetes.io/projected/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-kube-api-access-xg4vc\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239549 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-scripts\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239584 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s9p9\" (UniqueName: \"kubernetes.io/projected/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-kube-api-access-2s9p9\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239603 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239620 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-config-data\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239638 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239658 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-logs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239685 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239704 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239753 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239772 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-config-data\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239788 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239815 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-scripts\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239835 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-logs\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239853 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239876 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.239894 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.240720 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.242349 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.245674 4756 scope.go:117] "RemoveContainer" containerID="910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.247420 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.247799 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-logs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.249439 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.252763 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-logs\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.258989 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-config-data\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.271511 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.284172 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.284647 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.284972 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-scripts\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.286621 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.301229 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s9p9\" (UniqueName: \"kubernetes.io/projected/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-kube-api-access-2s9p9\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.302212 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-config-data\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.302311 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.303808 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a921c4c-56ce-4689-b25d-0b4b69a01cdb-scripts\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.307026 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg4vc\" (UniqueName: \"kubernetes.io/projected/1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26-kube-api-access-xg4vc\") pod \"cinder-api-0\" (UID: \"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26\") " pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.331030 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.343773 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2a921c4c-56ce-4689-b25d-0b4b69a01cdb\") " pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.514240 4756 scope.go:117] "RemoveContainer" containerID="86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30" Feb 03 09:29:30 crc kubenswrapper[4756]: E0203 09:29:30.517325 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30\": container with ID starting with 86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30 not found: ID does not exist" containerID="86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.517359 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30"} err="failed to get container status \"86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30\": rpc error: code = NotFound desc = could not find container \"86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30\": container with ID starting with 86f2ba0f0631da9a45629cc2cd821cd63687a157801009c2f89c0550c9b56f30 not found: ID does not exist" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.517383 4756 scope.go:117] "RemoveContainer" containerID="910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0" Feb 03 09:29:30 crc kubenswrapper[4756]: E0203 09:29:30.517985 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0\": container with ID starting with 910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0 not found: ID does not exist" containerID="910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.518019 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0"} err="failed to get container status \"910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0\": rpc error: code = NotFound desc = could not find container \"910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0\": container with ID starting with 910903e53a8488d7d150375727d424bfb00b9aa61905ab465368bbb26518eef0 not found: ID does not exist" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.518254 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.529229 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.546656 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rbvb\" (UniqueName: \"kubernetes.io/projected/24066fdc-6998-4f09-9a07-230cb3378f11-kube-api-access-6rbvb\") pod \"24066fdc-6998-4f09-9a07-230cb3378f11\" (UID: \"24066fdc-6998-4f09-9a07-230cb3378f11\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.546956 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24066fdc-6998-4f09-9a07-230cb3378f11-operator-scripts\") pod \"24066fdc-6998-4f09-9a07-230cb3378f11\" (UID: \"24066fdc-6998-4f09-9a07-230cb3378f11\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.548163 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24066fdc-6998-4f09-9a07-230cb3378f11-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "24066fdc-6998-4f09-9a07-230cb3378f11" (UID: "24066fdc-6998-4f09-9a07-230cb3378f11"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.557049 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24066fdc-6998-4f09-9a07-230cb3378f11-kube-api-access-6rbvb" (OuterVolumeSpecName: "kube-api-access-6rbvb") pod "24066fdc-6998-4f09-9a07-230cb3378f11" (UID: "24066fdc-6998-4f09-9a07-230cb3378f11"). InnerVolumeSpecName "kube-api-access-6rbvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.575076 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.648338 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.649067 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b922d8-73c3-42e7-a7af-f28120f67125-operator-scripts\") pod \"a2b922d8-73c3-42e7-a7af-f28120f67125\" (UID: \"a2b922d8-73c3-42e7-a7af-f28120f67125\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.649231 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgt5k\" (UniqueName: \"kubernetes.io/projected/a2b922d8-73c3-42e7-a7af-f28120f67125-kube-api-access-lgt5k\") pod \"a2b922d8-73c3-42e7-a7af-f28120f67125\" (UID: \"a2b922d8-73c3-42e7-a7af-f28120f67125\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.651940 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2b922d8-73c3-42e7-a7af-f28120f67125-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a2b922d8-73c3-42e7-a7af-f28120f67125" (UID: "a2b922d8-73c3-42e7-a7af-f28120f67125"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.653949 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24066fdc-6998-4f09-9a07-230cb3378f11-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.653978 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2b922d8-73c3-42e7-a7af-f28120f67125-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.653991 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rbvb\" (UniqueName: \"kubernetes.io/projected/24066fdc-6998-4f09-9a07-230cb3378f11-kube-api-access-6rbvb\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.657898 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b922d8-73c3-42e7-a7af-f28120f67125-kube-api-access-lgt5k" (OuterVolumeSpecName: "kube-api-access-lgt5k") pod "a2b922d8-73c3-42e7-a7af-f28120f67125" (UID: "a2b922d8-73c3-42e7-a7af-f28120f67125"). InnerVolumeSpecName "kube-api-access-lgt5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.725693 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.736712 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.765041 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a12c2e-5874-45c7-bc27-1281125fb024-operator-scripts\") pod \"15a12c2e-5874-45c7-bc27-1281125fb024\" (UID: \"15a12c2e-5874-45c7-bc27-1281125fb024\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.765098 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdpwn\" (UniqueName: \"kubernetes.io/projected/4172aecf-8ba4-4990-9bd9-11ef49b0320e-kube-api-access-bdpwn\") pod \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\" (UID: \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.765131 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4172aecf-8ba4-4990-9bd9-11ef49b0320e-operator-scripts\") pod \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\" (UID: \"4172aecf-8ba4-4990-9bd9-11ef49b0320e\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.765205 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-operator-scripts\") pod \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\" (UID: \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.765271 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm486\" (UniqueName: \"kubernetes.io/projected/15a12c2e-5874-45c7-bc27-1281125fb024-kube-api-access-rm486\") pod \"15a12c2e-5874-45c7-bc27-1281125fb024\" (UID: \"15a12c2e-5874-45c7-bc27-1281125fb024\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.765406 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7dgf\" (UniqueName: \"kubernetes.io/projected/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-kube-api-access-x7dgf\") pod \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\" (UID: \"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.765842 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgt5k\" (UniqueName: \"kubernetes.io/projected/a2b922d8-73c3-42e7-a7af-f28120f67125-kube-api-access-lgt5k\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.767721 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15a12c2e-5874-45c7-bc27-1281125fb024-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15a12c2e-5874-45c7-bc27-1281125fb024" (UID: "15a12c2e-5874-45c7-bc27-1281125fb024"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.769024 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea" (UID: "d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.780057 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4172aecf-8ba4-4990-9bd9-11ef49b0320e-kube-api-access-bdpwn" (OuterVolumeSpecName: "kube-api-access-bdpwn") pod "4172aecf-8ba4-4990-9bd9-11ef49b0320e" (UID: "4172aecf-8ba4-4990-9bd9-11ef49b0320e"). InnerVolumeSpecName "kube-api-access-bdpwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.785844 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.786780 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4172aecf-8ba4-4990-9bd9-11ef49b0320e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4172aecf-8ba4-4990-9bd9-11ef49b0320e" (UID: "4172aecf-8ba4-4990-9bd9-11ef49b0320e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.790846 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a12c2e-5874-45c7-bc27-1281125fb024-kube-api-access-rm486" (OuterVolumeSpecName: "kube-api-access-rm486") pod "15a12c2e-5874-45c7-bc27-1281125fb024" (UID: "15a12c2e-5874-45c7-bc27-1281125fb024"). InnerVolumeSpecName "kube-api-access-rm486". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.792220 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-kube-api-access-x7dgf" (OuterVolumeSpecName: "kube-api-access-x7dgf") pod "d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea" (UID: "d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea"). InnerVolumeSpecName "kube-api-access-x7dgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.833091 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.845362 4756 generic.go:334] "Generic (PLEG): container finished" podID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerID="5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b" exitCode=0 Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.845459 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16c4732e-295a-495d-a2af-e10b2c71a10f","Type":"ContainerDied","Data":"5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b"} Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.845486 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16c4732e-295a-495d-a2af-e10b2c71a10f","Type":"ContainerDied","Data":"9edfe39d891b6ce82ffbe58805e9ccfb660934332dce6f97d42ae01582069c23"} Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.845503 4756 scope.go:117] "RemoveContainer" containerID="5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.845658 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.866310 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d2m25" event={"ID":"15a12c2e-5874-45c7-bc27-1281125fb024","Type":"ContainerDied","Data":"e14b8f867734a8d4fb06a3744eb7e3c6b016cc0714f0b81789b977e160d20ce4"} Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.866530 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e14b8f867734a8d4fb06a3744eb7e3c6b016cc0714f0b81789b977e160d20ce4" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.866661 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d2m25" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869199 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89hrk\" (UniqueName: \"kubernetes.io/projected/95eac2ad-73dc-405d-b583-92d263bec833-kube-api-access-89hrk\") pod \"95eac2ad-73dc-405d-b583-92d263bec833\" (UID: \"95eac2ad-73dc-405d-b583-92d263bec833\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869256 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"16c4732e-295a-495d-a2af-e10b2c71a10f\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869299 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-scripts\") pod \"16c4732e-295a-495d-a2af-e10b2c71a10f\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869324 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-internal-tls-certs\") pod \"16c4732e-295a-495d-a2af-e10b2c71a10f\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869344 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-combined-ca-bundle\") pod \"16c4732e-295a-495d-a2af-e10b2c71a10f\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869372 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnknm\" (UniqueName: \"kubernetes.io/projected/16c4732e-295a-495d-a2af-e10b2c71a10f-kube-api-access-wnknm\") pod \"16c4732e-295a-495d-a2af-e10b2c71a10f\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869426 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95eac2ad-73dc-405d-b583-92d263bec833-operator-scripts\") pod \"95eac2ad-73dc-405d-b583-92d263bec833\" (UID: \"95eac2ad-73dc-405d-b583-92d263bec833\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869513 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-config-data\") pod \"16c4732e-295a-495d-a2af-e10b2c71a10f\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869549 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-logs\") pod \"16c4732e-295a-495d-a2af-e10b2c71a10f\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869579 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-httpd-run\") pod \"16c4732e-295a-495d-a2af-e10b2c71a10f\" (UID: \"16c4732e-295a-495d-a2af-e10b2c71a10f\") " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869931 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm486\" (UniqueName: \"kubernetes.io/projected/15a12c2e-5874-45c7-bc27-1281125fb024-kube-api-access-rm486\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869941 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7dgf\" (UniqueName: \"kubernetes.io/projected/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-kube-api-access-x7dgf\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869950 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a12c2e-5874-45c7-bc27-1281125fb024-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869959 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdpwn\" (UniqueName: \"kubernetes.io/projected/4172aecf-8ba4-4990-9bd9-11ef49b0320e-kube-api-access-bdpwn\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869966 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4172aecf-8ba4-4990-9bd9-11ef49b0320e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.869974 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.870296 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "16c4732e-295a-495d-a2af-e10b2c71a10f" (UID: "16c4732e-295a-495d-a2af-e10b2c71a10f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.872817 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95eac2ad-73dc-405d-b583-92d263bec833-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95eac2ad-73dc-405d-b583-92d263bec833" (UID: "95eac2ad-73dc-405d-b583-92d263bec833"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.874784 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-logs" (OuterVolumeSpecName: "logs") pod "16c4732e-295a-495d-a2af-e10b2c71a10f" (UID: "16c4732e-295a-495d-a2af-e10b2c71a10f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.884795 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c4732e-295a-495d-a2af-e10b2c71a10f-kube-api-access-wnknm" (OuterVolumeSpecName: "kube-api-access-wnknm") pod "16c4732e-295a-495d-a2af-e10b2c71a10f" (UID: "16c4732e-295a-495d-a2af-e10b2c71a10f"). InnerVolumeSpecName "kube-api-access-wnknm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.887205 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95eac2ad-73dc-405d-b583-92d263bec833-kube-api-access-89hrk" (OuterVolumeSpecName: "kube-api-access-89hrk") pod "95eac2ad-73dc-405d-b583-92d263bec833" (UID: "95eac2ad-73dc-405d-b583-92d263bec833"). InnerVolumeSpecName "kube-api-access-89hrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.901340 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" event={"ID":"d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea","Type":"ContainerDied","Data":"1d83dd66fedd7334842056b52516c8e7a755fda81e9de4a297c99ee357496050"} Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.901389 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d83dd66fedd7334842056b52516c8e7a755fda81e9de4a297c99ee357496050" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.901470 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f9ed-account-create-update-jnkxm" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.915092 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "16c4732e-295a-495d-a2af-e10b2c71a10f" (UID: "16c4732e-295a-495d-a2af-e10b2c71a10f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.916617 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-scripts" (OuterVolumeSpecName: "scripts") pod "16c4732e-295a-495d-a2af-e10b2c71a10f" (UID: "16c4732e-295a-495d-a2af-e10b2c71a10f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.972303 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.972342 4756 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16c4732e-295a-495d-a2af-e10b2c71a10f-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.972355 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89hrk\" (UniqueName: \"kubernetes.io/projected/95eac2ad-73dc-405d-b583-92d263bec833-kube-api-access-89hrk\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.972376 4756 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.972385 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.972394 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnknm\" (UniqueName: \"kubernetes.io/projected/16c4732e-295a-495d-a2af-e10b2c71a10f-kube-api-access-wnknm\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.972403 4756 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95eac2ad-73dc-405d-b583-92d263bec833-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.976834 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16c4732e-295a-495d-a2af-e10b2c71a10f" (UID: "16c4732e-295a-495d-a2af-e10b2c71a10f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.980659 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "16c4732e-295a-495d-a2af-e10b2c71a10f" (UID: "16c4732e-295a-495d-a2af-e10b2c71a10f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.983791 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5fd1-account-create-update-6kqzn" event={"ID":"4172aecf-8ba4-4990-9bd9-11ef49b0320e","Type":"ContainerDied","Data":"0d906a731ab533d2aa77d84e5140bd916f0dbeb31f9e0ab29e7bc94e1bf134cb"} Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.983837 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d906a731ab533d2aa77d84e5140bd916f0dbeb31f9e0ab29e7bc94e1bf134cb" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.983871 4756 scope.go:117] "RemoveContainer" containerID="0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef" Feb 03 09:29:30 crc kubenswrapper[4756]: I0203 09:29:30.983996 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5fd1-account-create-update-6kqzn" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.001994 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-x676l" event={"ID":"95eac2ad-73dc-405d-b583-92d263bec833","Type":"ContainerDied","Data":"0d458f75237d04c9dc0d95e438455e4cf81ccc5faf301ea1f1715d7ac55b2794"} Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.002033 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d458f75237d04c9dc0d95e438455e4cf81ccc5faf301ea1f1715d7ac55b2794" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.002081 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-x676l" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.015767 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6s59j" event={"ID":"24066fdc-6998-4f09-9a07-230cb3378f11","Type":"ContainerDied","Data":"c57ab2175efbd77ee462bd3480ce8f25adb9cec38e7cf75dd0aadb3770fceb60"} Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.015805 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c57ab2175efbd77ee462bd3480ce8f25adb9cec38e7cf75dd0aadb3770fceb60" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.015861 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6s59j" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.025568 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerStarted","Data":"18a1fa4d8a0e19c6ce4bb001a39b44f449d70c3d413770a3888e3fd9a4846b73"} Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.031506 4756 scope.go:117] "RemoveContainer" containerID="5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.040035 4756 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.042833 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b\": container with ID starting with 5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b not found: ID does not exist" containerID="5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.042887 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b"} err="failed to get container status \"5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b\": rpc error: code = NotFound desc = could not find container \"5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b\": container with ID starting with 5570c44c9f076227014b757af362df4d9228659a82b7d1ad8d7fc42a9d1a364b not found: ID does not exist" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.043156 4756 scope.go:117] "RemoveContainer" containerID="0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.044724 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef\": container with ID starting with 0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef not found: ID does not exist" containerID="0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.044757 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef"} err="failed to get container status \"0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef\": rpc error: code = NotFound desc = could not find container \"0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef\": container with ID starting with 0668b82821c0313dd7031e3f11869d25a66773bff457fb57e20ce0be96c305ef not found: ID does not exist" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.045939 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f912-account-create-update-t2sfb" event={"ID":"a2b922d8-73c3-42e7-a7af-f28120f67125","Type":"ContainerDied","Data":"6d5696e7e9b8f6bd6f7370707e315eb4d2210eae0ea148376b7d4a4eaa00ef57"} Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.045970 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d5696e7e9b8f6bd6f7370707e315eb4d2210eae0ea148376b7d4a4eaa00ef57" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.046031 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f912-account-create-update-t2sfb" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.069431 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-config-data" (OuterVolumeSpecName: "config-data") pod "16c4732e-295a-495d-a2af-e10b2c71a10f" (UID: "16c4732e-295a-495d-a2af-e10b2c71a10f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.075528 4756 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.075560 4756 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.075571 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.075582 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c4732e-295a-495d-a2af-e10b2c71a10f-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.251506 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.268352 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.291684 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.292154 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-log" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292176 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-log" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.292199 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292208 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.292219 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a12c2e-5874-45c7-bc27-1281125fb024" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292226 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a12c2e-5874-45c7-bc27-1281125fb024" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.292237 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-httpd" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292246 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-httpd" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.292255 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4172aecf-8ba4-4990-9bd9-11ef49b0320e" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292263 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="4172aecf-8ba4-4990-9bd9-11ef49b0320e" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.292297 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b922d8-73c3-42e7-a7af-f28120f67125" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292304 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b922d8-73c3-42e7-a7af-f28120f67125" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.292315 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95eac2ad-73dc-405d-b583-92d263bec833" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292322 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="95eac2ad-73dc-405d-b583-92d263bec833" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: E0203 09:29:31.292333 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24066fdc-6998-4f09-9a07-230cb3378f11" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292340 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="24066fdc-6998-4f09-9a07-230cb3378f11" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292548 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="95eac2ad-73dc-405d-b583-92d263bec833" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292564 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-log" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292573 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" containerName="glance-httpd" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292583 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="24066fdc-6998-4f09-9a07-230cb3378f11" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292596 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="4172aecf-8ba4-4990-9bd9-11ef49b0320e" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292611 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b922d8-73c3-42e7-a7af-f28120f67125" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292623 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea" containerName="mariadb-account-create-update" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.292640 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a12c2e-5874-45c7-bc27-1281125fb024" containerName="mariadb-database-create" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.293853 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.298778 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.302823 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.341778 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.372565 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.389836 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd8465a-e888-4b54-a570-76bcdc4736ac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.389938 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf6dj\" (UniqueName: \"kubernetes.io/projected/2bd8465a-e888-4b54-a570-76bcdc4736ac-kube-api-access-bf6dj\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.390014 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.390113 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.390170 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd8465a-e888-4b54-a570-76bcdc4736ac-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.390214 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.390254 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.390297 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.492621 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf6dj\" (UniqueName: \"kubernetes.io/projected/2bd8465a-e888-4b54-a570-76bcdc4736ac-kube-api-access-bf6dj\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.492684 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.493307 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.493340 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd8465a-e888-4b54-a570-76bcdc4736ac-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.493622 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.493654 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.493677 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.493701 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd8465a-e888-4b54-a570-76bcdc4736ac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.493965 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd8465a-e888-4b54-a570-76bcdc4736ac-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.495969 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.498165 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.498357 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bd8465a-e888-4b54-a570-76bcdc4736ac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.501764 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.502140 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.513140 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf6dj\" (UniqueName: \"kubernetes.io/projected/2bd8465a-e888-4b54-a570-76bcdc4736ac-kube-api-access-bf6dj\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.538415 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd8465a-e888-4b54-a570-76bcdc4736ac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.563988 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bd8465a-e888-4b54-a570-76bcdc4736ac\") " pod="openstack/glance-default-internal-api-0" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.602753 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.631886 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16c4732e-295a-495d-a2af-e10b2c71a10f" path="/var/lib/kubelet/pods/16c4732e-295a-495d-a2af-e10b2c71a10f/volumes" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.633119 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ccb5fd-4e2d-4cc8-b776-161687b99eaa" path="/var/lib/kubelet/pods/19ccb5fd-4e2d-4cc8-b776-161687b99eaa/volumes" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.634323 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd03b9b6-ff6d-4eed-be6d-e08461b6c693" path="/var/lib/kubelet/pods/fd03b9b6-ff6d-4eed-be6d-e08461b6c693/volumes" Feb 03 09:29:31 crc kubenswrapper[4756]: I0203 09:29:31.673861 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.093429 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26","Type":"ContainerStarted","Data":"672bfdcbee801dd585ee9cd57b1f98868bfd72eee273ba3700b35f51089c1b8d"} Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.118801 4756 generic.go:334] "Generic (PLEG): container finished" podID="c571905e-5655-45c7-a864-70d02e7a9326" containerID="1b0973a2612cd6f68029aef4d5a1476dd1e24253e75900f5476116c0a5b7eca0" exitCode=0 Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.118877 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b74899c58-bsxwf" event={"ID":"c571905e-5655-45c7-a864-70d02e7a9326","Type":"ContainerDied","Data":"1b0973a2612cd6f68029aef4d5a1476dd1e24253e75900f5476116c0a5b7eca0"} Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.130640 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2a921c4c-56ce-4689-b25d-0b4b69a01cdb","Type":"ContainerStarted","Data":"dbfcfb7659e5a7d22f831f6837ba4f10f3c17266236b671f9951f4b4189d3ad9"} Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.309297 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.527608 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.619407 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-config\") pod \"c571905e-5655-45c7-a864-70d02e7a9326\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.619564 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds97c\" (UniqueName: \"kubernetes.io/projected/c571905e-5655-45c7-a864-70d02e7a9326-kube-api-access-ds97c\") pod \"c571905e-5655-45c7-a864-70d02e7a9326\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.619624 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-httpd-config\") pod \"c571905e-5655-45c7-a864-70d02e7a9326\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.619931 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-ovndb-tls-certs\") pod \"c571905e-5655-45c7-a864-70d02e7a9326\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.619985 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-combined-ca-bundle\") pod \"c571905e-5655-45c7-a864-70d02e7a9326\" (UID: \"c571905e-5655-45c7-a864-70d02e7a9326\") " Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.644562 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c571905e-5655-45c7-a864-70d02e7a9326" (UID: "c571905e-5655-45c7-a864-70d02e7a9326"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.648564 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c571905e-5655-45c7-a864-70d02e7a9326-kube-api-access-ds97c" (OuterVolumeSpecName: "kube-api-access-ds97c") pod "c571905e-5655-45c7-a864-70d02e7a9326" (UID: "c571905e-5655-45c7-a864-70d02e7a9326"). InnerVolumeSpecName "kube-api-access-ds97c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.728030 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds97c\" (UniqueName: \"kubernetes.io/projected/c571905e-5655-45c7-a864-70d02e7a9326-kube-api-access-ds97c\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.728338 4756 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.773294 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-config" (OuterVolumeSpecName: "config") pod "c571905e-5655-45c7-a864-70d02e7a9326" (UID: "c571905e-5655-45c7-a864-70d02e7a9326"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.788714 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c571905e-5655-45c7-a864-70d02e7a9326" (UID: "c571905e-5655-45c7-a864-70d02e7a9326"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.823646 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c571905e-5655-45c7-a864-70d02e7a9326" (UID: "c571905e-5655-45c7-a864-70d02e7a9326"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.833049 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.833258 4756 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:32 crc kubenswrapper[4756]: I0203 09:29:32.833353 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c571905e-5655-45c7-a864-70d02e7a9326-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.157239 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2a921c4c-56ce-4689-b25d-0b4b69a01cdb","Type":"ContainerStarted","Data":"e99e87470edf4dbac072af431784d65f872e50bfcbd80d2c34c473d25a81d7eb"} Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.159370 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerStarted","Data":"89e30e6e3cbb4c978aca36fed12ee6edf3990818c07bf3f57973f1831ed38f26"} Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.160351 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="ceilometer-central-agent" containerID="cri-o://f23a5200fcfb1ed2593606e2a184dd0f9adaade347b226fcad6da01d47f764b4" gracePeriod=30 Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.160582 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.160601 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="proxy-httpd" containerID="cri-o://89e30e6e3cbb4c978aca36fed12ee6edf3990818c07bf3f57973f1831ed38f26" gracePeriod=30 Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.160569 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="sg-core" containerID="cri-o://18a1fa4d8a0e19c6ce4bb001a39b44f449d70c3d413770a3888e3fd9a4846b73" gracePeriod=30 Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.160705 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="ceilometer-notification-agent" containerID="cri-o://2e54bad7a03e5dc604442aba7be16afeb3ed24cbd85f34680f4cce5c88096e95" gracePeriod=30 Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.176335 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26","Type":"ContainerStarted","Data":"b58586515ee6f49c83cb459015e39cc0654dd557e14fa038997fd0f2b2b665d2"} Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.179600 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd8465a-e888-4b54-a570-76bcdc4736ac","Type":"ContainerStarted","Data":"910778b945f63e5bacfa1f6066115e86cf4e8cf4492cff1c730dc2419df6877a"} Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.179630 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd8465a-e888-4b54-a570-76bcdc4736ac","Type":"ContainerStarted","Data":"8ab831c6ef420f550efc59b33bdedc6cf8ac6fd450f8924a56dafff4bbdf996b"} Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.184675 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b74899c58-bsxwf" event={"ID":"c571905e-5655-45c7-a864-70d02e7a9326","Type":"ContainerDied","Data":"6bcf1ad35168003f3d1acfc0ecc2eb531505d0eaee0b45e4a7011b050f8968f6"} Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.184710 4756 scope.go:117] "RemoveContainer" containerID="dedf5fd2bb1e43cfe9147a8997060b222a345753833e11dab8117b21d7572e1d" Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.184828 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b74899c58-bsxwf" Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.197475 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.58896471 podStartE2EDuration="8.197433849s" podCreationTimestamp="2026-02-03 09:29:25 +0000 UTC" firstStartedPulling="2026-02-03 09:29:26.870665894 +0000 UTC m=+1158.021133269" lastFinishedPulling="2026-02-03 09:29:32.479135033 +0000 UTC m=+1163.629602408" observedRunningTime="2026-02-03 09:29:33.188937443 +0000 UTC m=+1164.339404818" watchObservedRunningTime="2026-02-03 09:29:33.197433849 +0000 UTC m=+1164.347901224" Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.227868 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7b74899c58-bsxwf"] Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.239892 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7b74899c58-bsxwf"] Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.370046 4756 scope.go:117] "RemoveContainer" containerID="1b0973a2612cd6f68029aef4d5a1476dd1e24253e75900f5476116c0a5b7eca0" Feb 03 09:29:33 crc kubenswrapper[4756]: I0203 09:29:33.628858 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c571905e-5655-45c7-a864-70d02e7a9326" path="/var/lib/kubelet/pods/c571905e-5655-45c7-a864-70d02e7a9326/volumes" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.200258 4756 generic.go:334] "Generic (PLEG): container finished" podID="e45bd2fd-140c-409a-9592-252f4723354e" containerID="798d8b64be1bedcbb904e87c52b8e23badc5c7864eb3968e0e041473058e8551" exitCode=137 Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.200486 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-546db499dd-c5f8j" event={"ID":"e45bd2fd-140c-409a-9592-252f4723354e","Type":"ContainerDied","Data":"798d8b64be1bedcbb904e87c52b8e23badc5c7864eb3968e0e041473058e8551"} Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.203918 4756 generic.go:334] "Generic (PLEG): container finished" podID="84b66417-49cf-498b-ab27-3a8872227641" containerID="89e30e6e3cbb4c978aca36fed12ee6edf3990818c07bf3f57973f1831ed38f26" exitCode=0 Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.203941 4756 generic.go:334] "Generic (PLEG): container finished" podID="84b66417-49cf-498b-ab27-3a8872227641" containerID="18a1fa4d8a0e19c6ce4bb001a39b44f449d70c3d413770a3888e3fd9a4846b73" exitCode=2 Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.203949 4756 generic.go:334] "Generic (PLEG): container finished" podID="84b66417-49cf-498b-ab27-3a8872227641" containerID="2e54bad7a03e5dc604442aba7be16afeb3ed24cbd85f34680f4cce5c88096e95" exitCode=0 Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.203980 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerDied","Data":"89e30e6e3cbb4c978aca36fed12ee6edf3990818c07bf3f57973f1831ed38f26"} Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.204005 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerDied","Data":"18a1fa4d8a0e19c6ce4bb001a39b44f449d70c3d413770a3888e3fd9a4846b73"} Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.204017 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerDied","Data":"2e54bad7a03e5dc604442aba7be16afeb3ed24cbd85f34680f4cce5c88096e95"} Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.205986 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26","Type":"ContainerStarted","Data":"179c6da3dcc1f6bdd49b380cad7bec895bb35e96c3e0251cc7cd4b6dfe49b74d"} Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.207179 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.212794 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bd8465a-e888-4b54-a570-76bcdc4736ac","Type":"ContainerStarted","Data":"5caec8687145efa95c080ff3640da17fc824480cac0b58cb1139cff266ac65a1"} Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.215687 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2a921c4c-56ce-4689-b25d-0b4b69a01cdb","Type":"ContainerStarted","Data":"87b57122071136109f619f552a08cb558c5f2d5d9422af880d64bc2599286854"} Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.230151 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.23013304 podStartE2EDuration="5.23013304s" podCreationTimestamp="2026-02-03 09:29:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:34.226409063 +0000 UTC m=+1165.376876438" watchObservedRunningTime="2026-02-03 09:29:34.23013304 +0000 UTC m=+1165.380600415" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.252975 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.252955113 podStartE2EDuration="3.252955113s" podCreationTimestamp="2026-02-03 09:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:34.248864426 +0000 UTC m=+1165.399331811" watchObservedRunningTime="2026-02-03 09:29:34.252955113 +0000 UTC m=+1165.403422488" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.284720 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.284704177 podStartE2EDuration="5.284704177s" podCreationTimestamp="2026-02-03 09:29:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:29:34.281549558 +0000 UTC m=+1165.432016933" watchObservedRunningTime="2026-02-03 09:29:34.284704177 +0000 UTC m=+1165.435171552" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.860470 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.970035 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.973177 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-85ff9cc7c5-ns7sz" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.988054 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45bd2fd-140c-409a-9592-252f4723354e-logs\") pod \"e45bd2fd-140c-409a-9592-252f4723354e\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.988111 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-tls-certs\") pod \"e45bd2fd-140c-409a-9592-252f4723354e\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.988161 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s442\" (UniqueName: \"kubernetes.io/projected/e45bd2fd-140c-409a-9592-252f4723354e-kube-api-access-5s442\") pod \"e45bd2fd-140c-409a-9592-252f4723354e\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.988235 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-secret-key\") pod \"e45bd2fd-140c-409a-9592-252f4723354e\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.988310 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-scripts\") pod \"e45bd2fd-140c-409a-9592-252f4723354e\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.988361 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-config-data\") pod \"e45bd2fd-140c-409a-9592-252f4723354e\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.988389 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-combined-ca-bundle\") pod \"e45bd2fd-140c-409a-9592-252f4723354e\" (UID: \"e45bd2fd-140c-409a-9592-252f4723354e\") " Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.988567 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e45bd2fd-140c-409a-9592-252f4723354e-logs" (OuterVolumeSpecName: "logs") pod "e45bd2fd-140c-409a-9592-252f4723354e" (UID: "e45bd2fd-140c-409a-9592-252f4723354e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:34 crc kubenswrapper[4756]: I0203 09:29:34.989024 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e45bd2fd-140c-409a-9592-252f4723354e-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.015257 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e45bd2fd-140c-409a-9592-252f4723354e-kube-api-access-5s442" (OuterVolumeSpecName: "kube-api-access-5s442") pod "e45bd2fd-140c-409a-9592-252f4723354e" (UID: "e45bd2fd-140c-409a-9592-252f4723354e"). InnerVolumeSpecName "kube-api-access-5s442". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.022303 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e45bd2fd-140c-409a-9592-252f4723354e" (UID: "e45bd2fd-140c-409a-9592-252f4723354e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.030266 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-scripts" (OuterVolumeSpecName: "scripts") pod "e45bd2fd-140c-409a-9592-252f4723354e" (UID: "e45bd2fd-140c-409a-9592-252f4723354e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.054280 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-config-data" (OuterVolumeSpecName: "config-data") pod "e45bd2fd-140c-409a-9592-252f4723354e" (UID: "e45bd2fd-140c-409a-9592-252f4723354e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.055115 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "e45bd2fd-140c-409a-9592-252f4723354e" (UID: "e45bd2fd-140c-409a-9592-252f4723354e"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.058828 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e45bd2fd-140c-409a-9592-252f4723354e" (UID: "e45bd2fd-140c-409a-9592-252f4723354e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.090883 4756 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.090917 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s442\" (UniqueName: \"kubernetes.io/projected/e45bd2fd-140c-409a-9592-252f4723354e-kube-api-access-5s442\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.091033 4756 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.091353 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.091381 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e45bd2fd-140c-409a-9592-252f4723354e-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.091395 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45bd2fd-140c-409a-9592-252f4723354e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.228665 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-546db499dd-c5f8j" event={"ID":"e45bd2fd-140c-409a-9592-252f4723354e","Type":"ContainerDied","Data":"b0fa5b1cbc3eeb016c58e31ffcbd8b47631bcf967d79d7e8aba7d7056b011bc2"} Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.228739 4756 scope.go:117] "RemoveContainer" containerID="4f5b4620a588f33eab38c31e48cf91bf8ea74628a6ed552028a79834e5626fdb" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.228696 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-546db499dd-c5f8j" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.265908 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-546db499dd-c5f8j"] Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.275376 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-546db499dd-c5f8j"] Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.390172 4756 scope.go:117] "RemoveContainer" containerID="798d8b64be1bedcbb904e87c52b8e23badc5c7864eb3968e0e041473058e8551" Feb 03 09:29:35 crc kubenswrapper[4756]: I0203 09:29:35.624268 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e45bd2fd-140c-409a-9592-252f4723354e" path="/var/lib/kubelet/pods/e45bd2fd-140c-409a-9592-252f4723354e/volumes" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.245338 4756 generic.go:334] "Generic (PLEG): container finished" podID="84b66417-49cf-498b-ab27-3a8872227641" containerID="f23a5200fcfb1ed2593606e2a184dd0f9adaade347b226fcad6da01d47f764b4" exitCode=0 Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.245411 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerDied","Data":"f23a5200fcfb1ed2593606e2a184dd0f9adaade347b226fcad6da01d47f764b4"} Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.246046 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84b66417-49cf-498b-ab27-3a8872227641","Type":"ContainerDied","Data":"64fd379a227ea71f5d7f14bfffe85f5c40be5977f76e093b024338b1c7c447db"} Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.246064 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64fd379a227ea71f5d7f14bfffe85f5c40be5977f76e093b024338b1c7c447db" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.298360 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.422022 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-run-httpd\") pod \"84b66417-49cf-498b-ab27-3a8872227641\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.422199 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-scripts\") pod \"84b66417-49cf-498b-ab27-3a8872227641\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.422379 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhvpx\" (UniqueName: \"kubernetes.io/projected/84b66417-49cf-498b-ab27-3a8872227641-kube-api-access-fhvpx\") pod \"84b66417-49cf-498b-ab27-3a8872227641\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.422397 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "84b66417-49cf-498b-ab27-3a8872227641" (UID: "84b66417-49cf-498b-ab27-3a8872227641"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.422487 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-log-httpd\") pod \"84b66417-49cf-498b-ab27-3a8872227641\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.422596 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-sg-core-conf-yaml\") pod \"84b66417-49cf-498b-ab27-3a8872227641\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.422635 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-combined-ca-bundle\") pod \"84b66417-49cf-498b-ab27-3a8872227641\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.422668 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-config-data\") pod \"84b66417-49cf-498b-ab27-3a8872227641\" (UID: \"84b66417-49cf-498b-ab27-3a8872227641\") " Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.423056 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "84b66417-49cf-498b-ab27-3a8872227641" (UID: "84b66417-49cf-498b-ab27-3a8872227641"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.423471 4756 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.423498 4756 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84b66417-49cf-498b-ab27-3a8872227641-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.430234 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84b66417-49cf-498b-ab27-3a8872227641-kube-api-access-fhvpx" (OuterVolumeSpecName: "kube-api-access-fhvpx") pod "84b66417-49cf-498b-ab27-3a8872227641" (UID: "84b66417-49cf-498b-ab27-3a8872227641"). InnerVolumeSpecName "kube-api-access-fhvpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.445373 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-scripts" (OuterVolumeSpecName: "scripts") pod "84b66417-49cf-498b-ab27-3a8872227641" (UID: "84b66417-49cf-498b-ab27-3a8872227641"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.459101 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "84b66417-49cf-498b-ab27-3a8872227641" (UID: "84b66417-49cf-498b-ab27-3a8872227641"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.504662 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84b66417-49cf-498b-ab27-3a8872227641" (UID: "84b66417-49cf-498b-ab27-3a8872227641"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.525387 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.525427 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhvpx\" (UniqueName: \"kubernetes.io/projected/84b66417-49cf-498b-ab27-3a8872227641-kube-api-access-fhvpx\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.525440 4756 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.525471 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.533698 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-config-data" (OuterVolumeSpecName: "config-data") pod "84b66417-49cf-498b-ab27-3a8872227641" (UID: "84b66417-49cf-498b-ab27-3a8872227641"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.626939 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84b66417-49cf-498b-ab27-3a8872227641-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689149 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zkffm"] Feb 03 09:29:36 crc kubenswrapper[4756]: E0203 09:29:36.689533 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="ceilometer-notification-agent" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689550 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="ceilometer-notification-agent" Feb 03 09:29:36 crc kubenswrapper[4756]: E0203 09:29:36.689561 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="ceilometer-central-agent" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689567 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="ceilometer-central-agent" Feb 03 09:29:36 crc kubenswrapper[4756]: E0203 09:29:36.689577 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="sg-core" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689583 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="sg-core" Feb 03 09:29:36 crc kubenswrapper[4756]: E0203 09:29:36.689593 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c571905e-5655-45c7-a864-70d02e7a9326" containerName="neutron-httpd" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689600 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c571905e-5655-45c7-a864-70d02e7a9326" containerName="neutron-httpd" Feb 03 09:29:36 crc kubenswrapper[4756]: E0203 09:29:36.689615 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="proxy-httpd" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689620 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="proxy-httpd" Feb 03 09:29:36 crc kubenswrapper[4756]: E0203 09:29:36.689636 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon-log" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689643 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon-log" Feb 03 09:29:36 crc kubenswrapper[4756]: E0203 09:29:36.689657 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c571905e-5655-45c7-a864-70d02e7a9326" containerName="neutron-api" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689663 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c571905e-5655-45c7-a864-70d02e7a9326" containerName="neutron-api" Feb 03 09:29:36 crc kubenswrapper[4756]: E0203 09:29:36.689670 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689682 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689838 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="ceilometer-notification-agent" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689850 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="ceilometer-central-agent" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689859 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="c571905e-5655-45c7-a864-70d02e7a9326" containerName="neutron-api" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689868 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689873 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="sg-core" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689883 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="e45bd2fd-140c-409a-9592-252f4723354e" containerName="horizon-log" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689891 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="c571905e-5655-45c7-a864-70d02e7a9326" containerName="neutron-httpd" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.689902 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b66417-49cf-498b-ab27-3a8872227641" containerName="proxy-httpd" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.690404 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.692172 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-sjb78" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.692234 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.693030 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.706114 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zkffm"] Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.831362 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.831661 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrw6k\" (UniqueName: \"kubernetes.io/projected/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-kube-api-access-jrw6k\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.831776 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-scripts\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.831844 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-config-data\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.937554 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrw6k\" (UniqueName: \"kubernetes.io/projected/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-kube-api-access-jrw6k\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.937929 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-scripts\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.937955 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-config-data\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.938023 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.947384 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.950416 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-scripts\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.951047 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-config-data\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:36 crc kubenswrapper[4756]: I0203 09:29:36.968001 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrw6k\" (UniqueName: \"kubernetes.io/projected/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-kube-api-access-jrw6k\") pod \"nova-cell0-conductor-db-sync-zkffm\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.020886 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.255747 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.310561 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.321397 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.338395 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.341694 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.348832 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.349023 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.349650 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.445321 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-log-httpd\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.445456 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-run-httpd\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.445486 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-scripts\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.445507 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j29j7\" (UniqueName: \"kubernetes.io/projected/b8cee259-a509-47ec-8abd-fb850e50cfb3-kube-api-access-j29j7\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.445531 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.445581 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.445653 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-config-data\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.479096 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zkffm"] Feb 03 09:29:37 crc kubenswrapper[4756]: W0203 09:29:37.481206 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a34d89c_812e_4f86_8ec1_6a38cfd710c8.slice/crio-092b152cca90e9fc470e3c78b14a24e6c9492c6a0bb229982577655c9a877d8b WatchSource:0}: Error finding container 092b152cca90e9fc470e3c78b14a24e6c9492c6a0bb229982577655c9a877d8b: Status 404 returned error can't find the container with id 092b152cca90e9fc470e3c78b14a24e6c9492c6a0bb229982577655c9a877d8b Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.547391 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-log-httpd\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.547790 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-run-httpd\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.547809 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j29j7\" (UniqueName: \"kubernetes.io/projected/b8cee259-a509-47ec-8abd-fb850e50cfb3-kube-api-access-j29j7\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.548127 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-run-httpd\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.548040 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-log-httpd\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.547826 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-scripts\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.548197 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.548893 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.549393 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-config-data\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.553313 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.553488 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-scripts\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.567009 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.570031 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-config-data\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.571125 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j29j7\" (UniqueName: \"kubernetes.io/projected/b8cee259-a509-47ec-8abd-fb850e50cfb3-kube-api-access-j29j7\") pod \"ceilometer-0\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " pod="openstack/ceilometer-0" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.632860 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84b66417-49cf-498b-ab27-3a8872227641" path="/var/lib/kubelet/pods/84b66417-49cf-498b-ab27-3a8872227641/volumes" Feb 03 09:29:37 crc kubenswrapper[4756]: I0203 09:29:37.672130 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:29:38 crc kubenswrapper[4756]: I0203 09:29:38.124263 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:38 crc kubenswrapper[4756]: W0203 09:29:38.131816 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8cee259_a509_47ec_8abd_fb850e50cfb3.slice/crio-bb3df286cc8930a7beef94f103e928bd8513ef5c5b1ec46212fb999fd49af0eb WatchSource:0}: Error finding container bb3df286cc8930a7beef94f103e928bd8513ef5c5b1ec46212fb999fd49af0eb: Status 404 returned error can't find the container with id bb3df286cc8930a7beef94f103e928bd8513ef5c5b1ec46212fb999fd49af0eb Feb 03 09:29:38 crc kubenswrapper[4756]: I0203 09:29:38.266728 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerStarted","Data":"bb3df286cc8930a7beef94f103e928bd8513ef5c5b1ec46212fb999fd49af0eb"} Feb 03 09:29:38 crc kubenswrapper[4756]: I0203 09:29:38.269829 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zkffm" event={"ID":"5a34d89c-812e-4f86-8ec1-6a38cfd710c8","Type":"ContainerStarted","Data":"092b152cca90e9fc470e3c78b14a24e6c9492c6a0bb229982577655c9a877d8b"} Feb 03 09:29:39 crc kubenswrapper[4756]: I0203 09:29:39.279947 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerStarted","Data":"fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06"} Feb 03 09:29:39 crc kubenswrapper[4756]: I0203 09:29:39.838591 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.059721 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6f5fdf688d-jsswq" Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.204200 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ccf85886-5xql4"] Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.204507 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-ccf85886-5xql4" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerName="placement-log" containerID="cri-o://c8a4e13ff5983b0dc98fd0638ead6bbbfcd0258604e9de29739a075045be5e83" gracePeriod=30 Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.204675 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-ccf85886-5xql4" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerName="placement-api" containerID="cri-o://e19a82576e77552afbabff92a0e0ebf8849f51dc4b7c82ea480989825c3b0709" gracePeriod=30 Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.380778 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerStarted","Data":"8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966"} Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.533661 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.534375 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.585979 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 03 09:29:40 crc kubenswrapper[4756]: I0203 09:29:40.600476 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.314180 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.404275 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerStarted","Data":"3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe"} Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.409488 4756 generic.go:334] "Generic (PLEG): container finished" podID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerID="c8a4e13ff5983b0dc98fd0638ead6bbbfcd0258604e9de29739a075045be5e83" exitCode=143 Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.409699 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ccf85886-5xql4" event={"ID":"73d9e20e-cd45-437f-a2b5-a82370ed1038","Type":"ContainerDied","Data":"c8a4e13ff5983b0dc98fd0638ead6bbbfcd0258604e9de29739a075045be5e83"} Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.409819 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.409842 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.675117 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.676146 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.718137 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:41 crc kubenswrapper[4756]: I0203 09:29:41.724374 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:42 crc kubenswrapper[4756]: I0203 09:29:42.421071 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:42 crc kubenswrapper[4756]: I0203 09:29:42.421494 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:42 crc kubenswrapper[4756]: I0203 09:29:42.627671 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 03 09:29:43 crc kubenswrapper[4756]: I0203 09:29:43.571611 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 03 09:29:43 crc kubenswrapper[4756]: I0203 09:29:43.571714 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:29:43 crc kubenswrapper[4756]: I0203 09:29:43.575900 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 03 09:29:44 crc kubenswrapper[4756]: I0203 09:29:44.445569 4756 generic.go:334] "Generic (PLEG): container finished" podID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerID="e19a82576e77552afbabff92a0e0ebf8849f51dc4b7c82ea480989825c3b0709" exitCode=0 Feb 03 09:29:44 crc kubenswrapper[4756]: I0203 09:29:44.446138 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ccf85886-5xql4" event={"ID":"73d9e20e-cd45-437f-a2b5-a82370ed1038","Type":"ContainerDied","Data":"e19a82576e77552afbabff92a0e0ebf8849f51dc4b7c82ea480989825c3b0709"} Feb 03 09:29:44 crc kubenswrapper[4756]: I0203 09:29:44.446216 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:29:44 crc kubenswrapper[4756]: I0203 09:29:44.446230 4756 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 03 09:29:44 crc kubenswrapper[4756]: I0203 09:29:44.776891 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:44 crc kubenswrapper[4756]: I0203 09:29:44.781916 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.315047 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ccf85886-5xql4" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.397323 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-scripts\") pod \"73d9e20e-cd45-437f-a2b5-a82370ed1038\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.397461 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-internal-tls-certs\") pod \"73d9e20e-cd45-437f-a2b5-a82370ed1038\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.397484 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73d9e20e-cd45-437f-a2b5-a82370ed1038-logs\") pod \"73d9e20e-cd45-437f-a2b5-a82370ed1038\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.397533 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-config-data\") pod \"73d9e20e-cd45-437f-a2b5-a82370ed1038\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.397588 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-public-tls-certs\") pod \"73d9e20e-cd45-437f-a2b5-a82370ed1038\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.397644 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-combined-ca-bundle\") pod \"73d9e20e-cd45-437f-a2b5-a82370ed1038\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.397666 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltcjc\" (UniqueName: \"kubernetes.io/projected/73d9e20e-cd45-437f-a2b5-a82370ed1038-kube-api-access-ltcjc\") pod \"73d9e20e-cd45-437f-a2b5-a82370ed1038\" (UID: \"73d9e20e-cd45-437f-a2b5-a82370ed1038\") " Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.397863 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73d9e20e-cd45-437f-a2b5-a82370ed1038-logs" (OuterVolumeSpecName: "logs") pod "73d9e20e-cd45-437f-a2b5-a82370ed1038" (UID: "73d9e20e-cd45-437f-a2b5-a82370ed1038"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.398293 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73d9e20e-cd45-437f-a2b5-a82370ed1038-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.417143 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-scripts" (OuterVolumeSpecName: "scripts") pod "73d9e20e-cd45-437f-a2b5-a82370ed1038" (UID: "73d9e20e-cd45-437f-a2b5-a82370ed1038"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.418294 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73d9e20e-cd45-437f-a2b5-a82370ed1038-kube-api-access-ltcjc" (OuterVolumeSpecName: "kube-api-access-ltcjc") pod "73d9e20e-cd45-437f-a2b5-a82370ed1038" (UID: "73d9e20e-cd45-437f-a2b5-a82370ed1038"). InnerVolumeSpecName "kube-api-access-ltcjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.471879 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-config-data" (OuterVolumeSpecName: "config-data") pod "73d9e20e-cd45-437f-a2b5-a82370ed1038" (UID: "73d9e20e-cd45-437f-a2b5-a82370ed1038"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.473479 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73d9e20e-cd45-437f-a2b5-a82370ed1038" (UID: "73d9e20e-cd45-437f-a2b5-a82370ed1038"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.491122 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zkffm" event={"ID":"5a34d89c-812e-4f86-8ec1-6a38cfd710c8","Type":"ContainerStarted","Data":"65183e148b041543009d86cf8844f0a62e8424d31830d5c112de1d6df413fccd"} Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.494871 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerStarted","Data":"d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a"} Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.494902 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="ceilometer-central-agent" containerID="cri-o://fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06" gracePeriod=30 Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.494930 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.494976 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="ceilometer-notification-agent" containerID="cri-o://8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966" gracePeriod=30 Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.495050 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="proxy-httpd" containerID="cri-o://d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a" gracePeriod=30 Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.495141 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="sg-core" containerID="cri-o://3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe" gracePeriod=30 Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.500817 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.500842 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.500853 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltcjc\" (UniqueName: \"kubernetes.io/projected/73d9e20e-cd45-437f-a2b5-a82370ed1038-kube-api-access-ltcjc\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.500862 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.507927 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ccf85886-5xql4" event={"ID":"73d9e20e-cd45-437f-a2b5-a82370ed1038","Type":"ContainerDied","Data":"6ae9623200e35a3e332b68086aaa16453aea8843aa4f57fb16eb7a86f24e33a9"} Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.507988 4756 scope.go:117] "RemoveContainer" containerID="e19a82576e77552afbabff92a0e0ebf8849f51dc4b7c82ea480989825c3b0709" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.508187 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ccf85886-5xql4" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.525020 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-zkffm" podStartSLOduration=1.7632537369999999 podStartE2EDuration="13.524994944s" podCreationTimestamp="2026-02-03 09:29:36 +0000 UTC" firstStartedPulling="2026-02-03 09:29:37.483201448 +0000 UTC m=+1168.633668813" lastFinishedPulling="2026-02-03 09:29:49.244942645 +0000 UTC m=+1180.395410020" observedRunningTime="2026-02-03 09:29:49.508098006 +0000 UTC m=+1180.658565381" watchObservedRunningTime="2026-02-03 09:29:49.524994944 +0000 UTC m=+1180.675462319" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.544126 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.47156287 podStartE2EDuration="12.544103592s" podCreationTimestamp="2026-02-03 09:29:37 +0000 UTC" firstStartedPulling="2026-02-03 09:29:38.136574433 +0000 UTC m=+1169.287041808" lastFinishedPulling="2026-02-03 09:29:49.209115155 +0000 UTC m=+1180.359582530" observedRunningTime="2026-02-03 09:29:49.528204315 +0000 UTC m=+1180.678671700" watchObservedRunningTime="2026-02-03 09:29:49.544103592 +0000 UTC m=+1180.694570967" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.544367 4756 scope.go:117] "RemoveContainer" containerID="c8a4e13ff5983b0dc98fd0638ead6bbbfcd0258604e9de29739a075045be5e83" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.549106 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "73d9e20e-cd45-437f-a2b5-a82370ed1038" (UID: "73d9e20e-cd45-437f-a2b5-a82370ed1038"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.558566 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "73d9e20e-cd45-437f-a2b5-a82370ed1038" (UID: "73d9e20e-cd45-437f-a2b5-a82370ed1038"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.602869 4756 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.602904 4756 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d9e20e-cd45-437f-a2b5-a82370ed1038-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.829431 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ccf85886-5xql4"] Feb 03 09:29:49 crc kubenswrapper[4756]: I0203 09:29:49.861767 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ccf85886-5xql4"] Feb 03 09:29:50 crc kubenswrapper[4756]: I0203 09:29:50.520233 4756 generic.go:334] "Generic (PLEG): container finished" podID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerID="3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe" exitCode=2 Feb 03 09:29:50 crc kubenswrapper[4756]: I0203 09:29:50.520264 4756 generic.go:334] "Generic (PLEG): container finished" podID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerID="8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966" exitCode=0 Feb 03 09:29:50 crc kubenswrapper[4756]: I0203 09:29:50.520305 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerDied","Data":"3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe"} Feb 03 09:29:50 crc kubenswrapper[4756]: I0203 09:29:50.520362 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerDied","Data":"8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966"} Feb 03 09:29:51 crc kubenswrapper[4756]: I0203 09:29:51.631378 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" path="/var/lib/kubelet/pods/73d9e20e-cd45-437f-a2b5-a82370ed1038/volumes" Feb 03 09:29:54 crc kubenswrapper[4756]: I0203 09:29:54.561043 4756 generic.go:334] "Generic (PLEG): container finished" podID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerID="fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06" exitCode=0 Feb 03 09:29:54 crc kubenswrapper[4756]: I0203 09:29:54.561419 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerDied","Data":"fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06"} Feb 03 09:29:59 crc kubenswrapper[4756]: I0203 09:29:59.601612 4756 generic.go:334] "Generic (PLEG): container finished" podID="5a34d89c-812e-4f86-8ec1-6a38cfd710c8" containerID="65183e148b041543009d86cf8844f0a62e8424d31830d5c112de1d6df413fccd" exitCode=0 Feb 03 09:29:59 crc kubenswrapper[4756]: I0203 09:29:59.602119 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zkffm" event={"ID":"5a34d89c-812e-4f86-8ec1-6a38cfd710c8","Type":"ContainerDied","Data":"65183e148b041543009d86cf8844f0a62e8424d31830d5c112de1d6df413fccd"} Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.158690 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp"] Feb 03 09:30:00 crc kubenswrapper[4756]: E0203 09:30:00.159184 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerName="placement-api" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.159206 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerName="placement-api" Feb 03 09:30:00 crc kubenswrapper[4756]: E0203 09:30:00.159226 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerName="placement-log" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.159240 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerName="placement-log" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.159502 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerName="placement-api" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.159544 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="73d9e20e-cd45-437f-a2b5-a82370ed1038" containerName="placement-log" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.160227 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.162111 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.165205 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.170993 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp"] Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.291420 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqxsq\" (UniqueName: \"kubernetes.io/projected/abe4717b-f308-4623-8be9-bffc1097815a-kube-api-access-sqxsq\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.291485 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe4717b-f308-4623-8be9-bffc1097815a-secret-volume\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.291552 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe4717b-f308-4623-8be9-bffc1097815a-config-volume\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.393546 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe4717b-f308-4623-8be9-bffc1097815a-config-volume\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.393693 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqxsq\" (UniqueName: \"kubernetes.io/projected/abe4717b-f308-4623-8be9-bffc1097815a-kube-api-access-sqxsq\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.393747 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe4717b-f308-4623-8be9-bffc1097815a-secret-volume\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.394682 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe4717b-f308-4623-8be9-bffc1097815a-config-volume\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.402814 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe4717b-f308-4623-8be9-bffc1097815a-secret-volume\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.410828 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqxsq\" (UniqueName: \"kubernetes.io/projected/abe4717b-f308-4623-8be9-bffc1097815a-kube-api-access-sqxsq\") pod \"collect-profiles-29501850-5m9bp\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.488938 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.956038 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp"] Feb 03 09:30:00 crc kubenswrapper[4756]: I0203 09:30:00.961257 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.109970 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-scripts\") pod \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.110151 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-config-data\") pod \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.110179 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-combined-ca-bundle\") pod \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.110241 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrw6k\" (UniqueName: \"kubernetes.io/projected/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-kube-api-access-jrw6k\") pod \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\" (UID: \"5a34d89c-812e-4f86-8ec1-6a38cfd710c8\") " Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.116095 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-kube-api-access-jrw6k" (OuterVolumeSpecName: "kube-api-access-jrw6k") pod "5a34d89c-812e-4f86-8ec1-6a38cfd710c8" (UID: "5a34d89c-812e-4f86-8ec1-6a38cfd710c8"). InnerVolumeSpecName "kube-api-access-jrw6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.116128 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-scripts" (OuterVolumeSpecName: "scripts") pod "5a34d89c-812e-4f86-8ec1-6a38cfd710c8" (UID: "5a34d89c-812e-4f86-8ec1-6a38cfd710c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.135426 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-config-data" (OuterVolumeSpecName: "config-data") pod "5a34d89c-812e-4f86-8ec1-6a38cfd710c8" (UID: "5a34d89c-812e-4f86-8ec1-6a38cfd710c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.138920 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a34d89c-812e-4f86-8ec1-6a38cfd710c8" (UID: "5a34d89c-812e-4f86-8ec1-6a38cfd710c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.213231 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.213943 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.214068 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrw6k\" (UniqueName: \"kubernetes.io/projected/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-kube-api-access-jrw6k\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.214153 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a34d89c-812e-4f86-8ec1-6a38cfd710c8-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.635865 4756 generic.go:334] "Generic (PLEG): container finished" podID="abe4717b-f308-4623-8be9-bffc1097815a" containerID="281354e82600aa1b0aecb578ffbf58ed7afd8326567ce5bab438bcf72c9dcdc2" exitCode=0 Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.639165 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" event={"ID":"abe4717b-f308-4623-8be9-bffc1097815a","Type":"ContainerDied","Data":"281354e82600aa1b0aecb578ffbf58ed7afd8326567ce5bab438bcf72c9dcdc2"} Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.646592 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" event={"ID":"abe4717b-f308-4623-8be9-bffc1097815a","Type":"ContainerStarted","Data":"fd7d184c4a2572604582674f5cc8fa565762bc540f5f8e04c8036d45afe203f2"} Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.646638 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zkffm" event={"ID":"5a34d89c-812e-4f86-8ec1-6a38cfd710c8","Type":"ContainerDied","Data":"092b152cca90e9fc470e3c78b14a24e6c9492c6a0bb229982577655c9a877d8b"} Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.646651 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="092b152cca90e9fc470e3c78b14a24e6c9492c6a0bb229982577655c9a877d8b" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.642641 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zkffm" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.718988 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 03 09:30:01 crc kubenswrapper[4756]: E0203 09:30:01.719320 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a34d89c-812e-4f86-8ec1-6a38cfd710c8" containerName="nova-cell0-conductor-db-sync" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.719337 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a34d89c-812e-4f86-8ec1-6a38cfd710c8" containerName="nova-cell0-conductor-db-sync" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.719541 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a34d89c-812e-4f86-8ec1-6a38cfd710c8" containerName="nova-cell0-conductor-db-sync" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.720057 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.726782 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.726897 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-sjb78" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.730741 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.823865 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84eaadd5-45a9-4055-a567-9e72fe3141b2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.824114 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84eaadd5-45a9-4055-a567-9e72fe3141b2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.824264 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvnk4\" (UniqueName: \"kubernetes.io/projected/84eaadd5-45a9-4055-a567-9e72fe3141b2-kube-api-access-dvnk4\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.926258 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84eaadd5-45a9-4055-a567-9e72fe3141b2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.926753 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84eaadd5-45a9-4055-a567-9e72fe3141b2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.926791 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvnk4\" (UniqueName: \"kubernetes.io/projected/84eaadd5-45a9-4055-a567-9e72fe3141b2-kube-api-access-dvnk4\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.935952 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84eaadd5-45a9-4055-a567-9e72fe3141b2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.936406 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84eaadd5-45a9-4055-a567-9e72fe3141b2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:01 crc kubenswrapper[4756]: I0203 09:30:01.948404 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvnk4\" (UniqueName: \"kubernetes.io/projected/84eaadd5-45a9-4055-a567-9e72fe3141b2-kube-api-access-dvnk4\") pod \"nova-cell0-conductor-0\" (UID: \"84eaadd5-45a9-4055-a567-9e72fe3141b2\") " pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:02 crc kubenswrapper[4756]: I0203 09:30:02.036245 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:02 crc kubenswrapper[4756]: I0203 09:30:02.461703 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 03 09:30:02 crc kubenswrapper[4756]: I0203 09:30:02.653239 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"84eaadd5-45a9-4055-a567-9e72fe3141b2","Type":"ContainerStarted","Data":"6f4db551c104b6ff60a3b948a5ca2c07e189e5e5db0b3595d7646e826f55aed0"} Feb 03 09:30:02 crc kubenswrapper[4756]: I0203 09:30:02.942467 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.049752 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqxsq\" (UniqueName: \"kubernetes.io/projected/abe4717b-f308-4623-8be9-bffc1097815a-kube-api-access-sqxsq\") pod \"abe4717b-f308-4623-8be9-bffc1097815a\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.049848 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe4717b-f308-4623-8be9-bffc1097815a-secret-volume\") pod \"abe4717b-f308-4623-8be9-bffc1097815a\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.049866 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe4717b-f308-4623-8be9-bffc1097815a-config-volume\") pod \"abe4717b-f308-4623-8be9-bffc1097815a\" (UID: \"abe4717b-f308-4623-8be9-bffc1097815a\") " Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.050673 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abe4717b-f308-4623-8be9-bffc1097815a-config-volume" (OuterVolumeSpecName: "config-volume") pod "abe4717b-f308-4623-8be9-bffc1097815a" (UID: "abe4717b-f308-4623-8be9-bffc1097815a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.053564 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe4717b-f308-4623-8be9-bffc1097815a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "abe4717b-f308-4623-8be9-bffc1097815a" (UID: "abe4717b-f308-4623-8be9-bffc1097815a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.053581 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe4717b-f308-4623-8be9-bffc1097815a-kube-api-access-sqxsq" (OuterVolumeSpecName: "kube-api-access-sqxsq") pod "abe4717b-f308-4623-8be9-bffc1097815a" (UID: "abe4717b-f308-4623-8be9-bffc1097815a"). InnerVolumeSpecName "kube-api-access-sqxsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.151798 4756 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe4717b-f308-4623-8be9-bffc1097815a-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.151836 4756 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe4717b-f308-4623-8be9-bffc1097815a-config-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.151846 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqxsq\" (UniqueName: \"kubernetes.io/projected/abe4717b-f308-4623-8be9-bffc1097815a-kube-api-access-sqxsq\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.664287 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"84eaadd5-45a9-4055-a567-9e72fe3141b2","Type":"ContainerStarted","Data":"2a60671d0b2858bb68c71accfc649ac87fb83b309a49873f3019e8852cfd1c29"} Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.665828 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.668474 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" event={"ID":"abe4717b-f308-4623-8be9-bffc1097815a","Type":"ContainerDied","Data":"fd7d184c4a2572604582674f5cc8fa565762bc540f5f8e04c8036d45afe203f2"} Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.668507 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd7d184c4a2572604582674f5cc8fa565762bc540f5f8e04c8036d45afe203f2" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.668554 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp" Feb 03 09:30:03 crc kubenswrapper[4756]: I0203 09:30:03.693547 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.69352432 podStartE2EDuration="2.69352432s" podCreationTimestamp="2026-02-03 09:30:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:03.686419809 +0000 UTC m=+1194.836887264" watchObservedRunningTime="2026-02-03 09:30:03.69352432 +0000 UTC m=+1194.843991705" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.062105 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.490773 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-7kx6w"] Feb 03 09:30:07 crc kubenswrapper[4756]: E0203 09:30:07.491234 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe4717b-f308-4623-8be9-bffc1097815a" containerName="collect-profiles" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.491258 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe4717b-f308-4623-8be9-bffc1097815a" containerName="collect-profiles" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.491457 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="abe4717b-f308-4623-8be9-bffc1097815a" containerName="collect-profiles" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.492016 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.493737 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.497640 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.501864 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-7kx6w"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.645687 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lvrg\" (UniqueName: \"kubernetes.io/projected/197062ac-2a3d-4b9b-a099-0d5843e4c58b-kube-api-access-5lvrg\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.645777 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-scripts\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.645851 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-config-data\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.645898 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.689528 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.690943 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.697904 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.698688 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.700283 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.728361 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.729465 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.737898 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.747263 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-scripts\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.747349 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-config-data\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.747398 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.747574 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lvrg\" (UniqueName: \"kubernetes.io/projected/197062ac-2a3d-4b9b-a099-0d5843e4c58b-kube-api-access-5lvrg\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.757936 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-config-data\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.759486 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-scripts\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.763277 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.787238 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lvrg\" (UniqueName: \"kubernetes.io/projected/197062ac-2a3d-4b9b-a099-0d5843e4c58b-kube-api-access-5lvrg\") pod \"nova-cell0-cell-mapping-7kx6w\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.791278 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.808588 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.850811 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-config-data\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.850911 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-config-data\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.850947 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.850995 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7n9q\" (UniqueName: \"kubernetes.io/projected/62679a7b-6a34-499e-b0c9-8a50b3afe493-kube-api-access-j7n9q\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.851061 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.851086 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksgtc\" (UniqueName: \"kubernetes.io/projected/83584248-5055-4851-8e03-e07f8eb0c306-kube-api-access-ksgtc\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.851144 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62679a7b-6a34-499e-b0c9-8a50b3afe493-logs\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.858620 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.860033 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.864265 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.887559 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952366 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952461 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-config-data\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952507 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952576 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-config-data\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952613 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952663 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7n9q\" (UniqueName: \"kubernetes.io/projected/62679a7b-6a34-499e-b0c9-8a50b3afe493-kube-api-access-j7n9q\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952719 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hzfq\" (UniqueName: \"kubernetes.io/projected/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-kube-api-access-4hzfq\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952762 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952784 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksgtc\" (UniqueName: \"kubernetes.io/projected/83584248-5055-4851-8e03-e07f8eb0c306-kube-api-access-ksgtc\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.952853 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62679a7b-6a34-499e-b0c9-8a50b3afe493-logs\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.953322 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62679a7b-6a34-499e-b0c9-8a50b3afe493-logs\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.961815 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-config-data\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.961919 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.963710 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.965124 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.970097 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-config-data\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.972707 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.977337 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7n9q\" (UniqueName: \"kubernetes.io/projected/62679a7b-6a34-499e-b0c9-8a50b3afe493-kube-api-access-j7n9q\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.984183 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.985974 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " pod="openstack/nova-api-0" Feb 03 09:30:07 crc kubenswrapper[4756]: I0203 09:30:07.997672 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksgtc\" (UniqueName: \"kubernetes.io/projected/83584248-5055-4851-8e03-e07f8eb0c306-kube-api-access-ksgtc\") pod \"nova-scheduler-0\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.030500 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.065157 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.102869 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-vp5rl"] Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.119183 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.123520 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.125063 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-config-data\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.125131 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cde1d7b5-d043-427a-86d8-471ebbd4fad4-logs\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.125159 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hzfq\" (UniqueName: \"kubernetes.io/projected/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-kube-api-access-4hzfq\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.125284 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmtgn\" (UniqueName: \"kubernetes.io/projected/cde1d7b5-d043-427a-86d8-471ebbd4fad4-kube-api-access-lmtgn\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.125327 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.125384 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.134980 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-vp5rl"] Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.143217 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.150806 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.176679 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hzfq\" (UniqueName: \"kubernetes.io/projected/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-kube-api-access-4hzfq\") pod \"nova-cell1-novncproxy-0\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.227061 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cde1d7b5-d043-427a-86d8-471ebbd4fad4-logs\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.227368 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmtgn\" (UniqueName: \"kubernetes.io/projected/cde1d7b5-d043-427a-86d8-471ebbd4fad4-kube-api-access-lmtgn\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.227521 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.227619 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-config-data\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.229193 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cde1d7b5-d043-427a-86d8-471ebbd4fad4-logs\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.241544 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.242228 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-config-data\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.244262 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmtgn\" (UniqueName: \"kubernetes.io/projected/cde1d7b5-d043-427a-86d8-471ebbd4fad4-kube-api-access-lmtgn\") pod \"nova-metadata-0\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.325938 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.328937 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.329020 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-config\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.329076 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.329119 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.329140 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-svc\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.329158 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz2j2\" (UniqueName: \"kubernetes.io/projected/f34a237b-60ac-4350-a1c3-6dc43b603fd7-kube-api-access-nz2j2\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.392325 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.431329 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.431410 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-config\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.431469 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.431526 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.431559 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-svc\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.431584 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz2j2\" (UniqueName: \"kubernetes.io/projected/f34a237b-60ac-4350-a1c3-6dc43b603fd7-kube-api-access-nz2j2\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.433257 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.433614 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-svc\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.433647 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.433719 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-config\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.433720 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.456810 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz2j2\" (UniqueName: \"kubernetes.io/projected/f34a237b-60ac-4350-a1c3-6dc43b603fd7-kube-api-access-nz2j2\") pod \"dnsmasq-dns-bccf8f775-vp5rl\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.486148 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.543434 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-7kx6w"] Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.730398 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.738545 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7kx6w" event={"ID":"197062ac-2a3d-4b9b-a099-0d5843e4c58b","Type":"ContainerStarted","Data":"e4095f1c727b4bf683211b46a74ab8b90e3b8f6fa9937e1e235aaf000b43d415"} Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.745234 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4qgbq"] Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.746813 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.748942 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.749841 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.756879 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4qgbq"] Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.849121 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:08 crc kubenswrapper[4756]: W0203 09:30:08.849386 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83584248_5055_4851_8e03_e07f8eb0c306.slice/crio-9e8e596f08f6597b6e3ba34cab1a3af5c3f8972cbf14c4a3298e986ad8624b7f WatchSource:0}: Error finding container 9e8e596f08f6597b6e3ba34cab1a3af5c3f8972cbf14c4a3298e986ad8624b7f: Status 404 returned error can't find the container with id 9e8e596f08f6597b6e3ba34cab1a3af5c3f8972cbf14c4a3298e986ad8624b7f Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.940816 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-config-data\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.941156 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.941226 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-scripts\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.941294 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mp27\" (UniqueName: \"kubernetes.io/projected/5af01732-2a75-4a7d-af19-b097bc0ffd6b-kube-api-access-6mp27\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:08 crc kubenswrapper[4756]: I0203 09:30:08.968715 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:08 crc kubenswrapper[4756]: W0203 09:30:08.972835 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f77a4df_b751_4e2e_a5bd_d22f4f453f81.slice/crio-30949ca8679459576e41152df1d0fc6d6ee30c53693c2182135e6fd50928824b WatchSource:0}: Error finding container 30949ca8679459576e41152df1d0fc6d6ee30c53693c2182135e6fd50928824b: Status 404 returned error can't find the container with id 30949ca8679459576e41152df1d0fc6d6ee30c53693c2182135e6fd50928824b Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.042759 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-config-data\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.042816 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.042872 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-scripts\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.042938 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mp27\" (UniqueName: \"kubernetes.io/projected/5af01732-2a75-4a7d-af19-b097bc0ffd6b-kube-api-access-6mp27\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.050203 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-scripts\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.050283 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-config-data\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.050601 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.064118 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mp27\" (UniqueName: \"kubernetes.io/projected/5af01732-2a75-4a7d-af19-b097bc0ffd6b-kube-api-access-6mp27\") pod \"nova-cell1-conductor-db-sync-4qgbq\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.071516 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.148622 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-vp5rl"] Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.196060 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.756585 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4qgbq"] Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.775863 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5f77a4df-b751-4e2e-a5bd-d22f4f453f81","Type":"ContainerStarted","Data":"30949ca8679459576e41152df1d0fc6d6ee30c53693c2182135e6fd50928824b"} Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.783605 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62679a7b-6a34-499e-b0c9-8a50b3afe493","Type":"ContainerStarted","Data":"fa77de810e7df0e67fcdbfb545de075bb1cf83ca87cce1f7198acef98627659d"} Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.788242 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83584248-5055-4851-8e03-e07f8eb0c306","Type":"ContainerStarted","Data":"9e8e596f08f6597b6e3ba34cab1a3af5c3f8972cbf14c4a3298e986ad8624b7f"} Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.795806 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" event={"ID":"5af01732-2a75-4a7d-af19-b097bc0ffd6b","Type":"ContainerStarted","Data":"78e0aaaa531bb43cff4e6bbb00a1d4cb32f170a056cc7450cbbfe3de01d3d90c"} Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.799151 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7kx6w" event={"ID":"197062ac-2a3d-4b9b-a099-0d5843e4c58b","Type":"ContainerStarted","Data":"23531e769d889687f55e0157730d72e4aaae9107e14ac8e940390a190adeafbe"} Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.802556 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cde1d7b5-d043-427a-86d8-471ebbd4fad4","Type":"ContainerStarted","Data":"97c8f45e69b8146a7f5f3c5e39692b0b85c7fc9251566cce210f753cf7dd8e02"} Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.808088 4756 generic.go:334] "Generic (PLEG): container finished" podID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" containerID="fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb" exitCode=0 Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.808161 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" event={"ID":"f34a237b-60ac-4350-a1c3-6dc43b603fd7","Type":"ContainerDied","Data":"fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb"} Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.808214 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" event={"ID":"f34a237b-60ac-4350-a1c3-6dc43b603fd7","Type":"ContainerStarted","Data":"02f338939498c8fd9041b5a032a6005ac3f42c1a01ce39c09d95dddba525890e"} Feb 03 09:30:09 crc kubenswrapper[4756]: I0203 09:30:09.877227 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-7kx6w" podStartSLOduration=2.8772081800000002 podStartE2EDuration="2.87720818s" podCreationTimestamp="2026-02-03 09:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:09.853365415 +0000 UTC m=+1201.003832790" watchObservedRunningTime="2026-02-03 09:30:09.87720818 +0000 UTC m=+1201.027675555" Feb 03 09:30:10 crc kubenswrapper[4756]: I0203 09:30:10.821903 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" event={"ID":"f34a237b-60ac-4350-a1c3-6dc43b603fd7","Type":"ContainerStarted","Data":"b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30"} Feb 03 09:30:10 crc kubenswrapper[4756]: I0203 09:30:10.823242 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:10 crc kubenswrapper[4756]: I0203 09:30:10.831791 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" event={"ID":"5af01732-2a75-4a7d-af19-b097bc0ffd6b","Type":"ContainerStarted","Data":"81c525f541f93ae9180d083bfaf1b9f216fb3827bccee3cccc00d04cebf6a0ca"} Feb 03 09:30:10 crc kubenswrapper[4756]: I0203 09:30:10.848811 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" podStartSLOduration=2.848795639 podStartE2EDuration="2.848795639s" podCreationTimestamp="2026-02-03 09:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:10.845986871 +0000 UTC m=+1201.996454246" watchObservedRunningTime="2026-02-03 09:30:10.848795639 +0000 UTC m=+1201.999263024" Feb 03 09:30:10 crc kubenswrapper[4756]: I0203 09:30:10.865368 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" podStartSLOduration=2.8653518780000002 podStartE2EDuration="2.865351878s" podCreationTimestamp="2026-02-03 09:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:10.861406474 +0000 UTC m=+1202.011873849" watchObservedRunningTime="2026-02-03 09:30:10.865351878 +0000 UTC m=+1202.015819253" Feb 03 09:30:11 crc kubenswrapper[4756]: I0203 09:30:11.501548 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:11 crc kubenswrapper[4756]: I0203 09:30:11.516859 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.852547 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62679a7b-6a34-499e-b0c9-8a50b3afe493","Type":"ContainerStarted","Data":"052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc"} Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.853048 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62679a7b-6a34-499e-b0c9-8a50b3afe493","Type":"ContainerStarted","Data":"d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf"} Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.855466 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83584248-5055-4851-8e03-e07f8eb0c306","Type":"ContainerStarted","Data":"2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f"} Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.858173 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cde1d7b5-d043-427a-86d8-471ebbd4fad4","Type":"ContainerStarted","Data":"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533"} Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.858205 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cde1d7b5-d043-427a-86d8-471ebbd4fad4","Type":"ContainerStarted","Data":"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c"} Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.858296 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerName="nova-metadata-log" containerID="cri-o://fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c" gracePeriod=30 Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.858566 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerName="nova-metadata-metadata" containerID="cri-o://f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533" gracePeriod=30 Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.861033 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5f77a4df-b751-4e2e-a5bd-d22f4f453f81" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2fe920db5b7ff2901fe8a303936aaa9ddb30bd04c5c80cf24c7208d76d56de96" gracePeriod=30 Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.861098 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5f77a4df-b751-4e2e-a5bd-d22f4f453f81","Type":"ContainerStarted","Data":"2fe920db5b7ff2901fe8a303936aaa9ddb30bd04c5c80cf24c7208d76d56de96"} Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.898284 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.572426637 podStartE2EDuration="5.898265241s" podCreationTimestamp="2026-02-03 09:30:07 +0000 UTC" firstStartedPulling="2026-02-03 09:30:08.736023407 +0000 UTC m=+1199.886490802" lastFinishedPulling="2026-02-03 09:30:12.061862031 +0000 UTC m=+1203.212329406" observedRunningTime="2026-02-03 09:30:12.872680462 +0000 UTC m=+1204.023147857" watchObservedRunningTime="2026-02-03 09:30:12.898265241 +0000 UTC m=+1204.048732616" Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.914052 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.716710181 podStartE2EDuration="5.913936362s" podCreationTimestamp="2026-02-03 09:30:07 +0000 UTC" firstStartedPulling="2026-02-03 09:30:08.860305815 +0000 UTC m=+1200.010773190" lastFinishedPulling="2026-02-03 09:30:12.057531996 +0000 UTC m=+1203.207999371" observedRunningTime="2026-02-03 09:30:12.889199358 +0000 UTC m=+1204.039666733" watchObservedRunningTime="2026-02-03 09:30:12.913936362 +0000 UTC m=+1204.064403737" Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.958604 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.879072419 podStartE2EDuration="5.958582998s" podCreationTimestamp="2026-02-03 09:30:07 +0000 UTC" firstStartedPulling="2026-02-03 09:30:08.97524417 +0000 UTC m=+1200.125711555" lastFinishedPulling="2026-02-03 09:30:12.054754759 +0000 UTC m=+1203.205222134" observedRunningTime="2026-02-03 09:30:12.911433614 +0000 UTC m=+1204.061900989" watchObservedRunningTime="2026-02-03 09:30:12.958582998 +0000 UTC m=+1204.109050383" Feb 03 09:30:12 crc kubenswrapper[4756]: I0203 09:30:12.966120 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.984118614 podStartE2EDuration="5.966103523s" podCreationTimestamp="2026-02-03 09:30:07 +0000 UTC" firstStartedPulling="2026-02-03 09:30:09.07533619 +0000 UTC m=+1200.225803565" lastFinishedPulling="2026-02-03 09:30:12.057321089 +0000 UTC m=+1203.207788474" observedRunningTime="2026-02-03 09:30:12.933654739 +0000 UTC m=+1204.084122144" watchObservedRunningTime="2026-02-03 09:30:12.966103523 +0000 UTC m=+1204.116570898" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.066209 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.326040 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.394181 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.394580 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.734144 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.859428 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmtgn\" (UniqueName: \"kubernetes.io/projected/cde1d7b5-d043-427a-86d8-471ebbd4fad4-kube-api-access-lmtgn\") pod \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.859658 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cde1d7b5-d043-427a-86d8-471ebbd4fad4-logs\") pod \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.859789 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-config-data\") pod \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.859827 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-combined-ca-bundle\") pod \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\" (UID: \"cde1d7b5-d043-427a-86d8-471ebbd4fad4\") " Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.859965 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cde1d7b5-d043-427a-86d8-471ebbd4fad4-logs" (OuterVolumeSpecName: "logs") pod "cde1d7b5-d043-427a-86d8-471ebbd4fad4" (UID: "cde1d7b5-d043-427a-86d8-471ebbd4fad4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.860432 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cde1d7b5-d043-427a-86d8-471ebbd4fad4-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.870131 4756 generic.go:334] "Generic (PLEG): container finished" podID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerID="f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533" exitCode=0 Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.870171 4756 generic.go:334] "Generic (PLEG): container finished" podID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerID="fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c" exitCode=143 Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.870704 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cde1d7b5-d043-427a-86d8-471ebbd4fad4-kube-api-access-lmtgn" (OuterVolumeSpecName: "kube-api-access-lmtgn") pod "cde1d7b5-d043-427a-86d8-471ebbd4fad4" (UID: "cde1d7b5-d043-427a-86d8-471ebbd4fad4"). InnerVolumeSpecName "kube-api-access-lmtgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.871041 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.871384 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cde1d7b5-d043-427a-86d8-471ebbd4fad4","Type":"ContainerDied","Data":"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533"} Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.871416 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cde1d7b5-d043-427a-86d8-471ebbd4fad4","Type":"ContainerDied","Data":"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c"} Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.871427 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cde1d7b5-d043-427a-86d8-471ebbd4fad4","Type":"ContainerDied","Data":"97c8f45e69b8146a7f5f3c5e39692b0b85c7fc9251566cce210f753cf7dd8e02"} Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.871454 4756 scope.go:117] "RemoveContainer" containerID="f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.885206 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-config-data" (OuterVolumeSpecName: "config-data") pod "cde1d7b5-d043-427a-86d8-471ebbd4fad4" (UID: "cde1d7b5-d043-427a-86d8-471ebbd4fad4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.889753 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cde1d7b5-d043-427a-86d8-471ebbd4fad4" (UID: "cde1d7b5-d043-427a-86d8-471ebbd4fad4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.925696 4756 scope.go:117] "RemoveContainer" containerID="fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.943276 4756 scope.go:117] "RemoveContainer" containerID="f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533" Feb 03 09:30:13 crc kubenswrapper[4756]: E0203 09:30:13.943765 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533\": container with ID starting with f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533 not found: ID does not exist" containerID="f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.943801 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533"} err="failed to get container status \"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533\": rpc error: code = NotFound desc = could not find container \"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533\": container with ID starting with f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533 not found: ID does not exist" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.943827 4756 scope.go:117] "RemoveContainer" containerID="fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c" Feb 03 09:30:13 crc kubenswrapper[4756]: E0203 09:30:13.944190 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c\": container with ID starting with fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c not found: ID does not exist" containerID="fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.944211 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c"} err="failed to get container status \"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c\": rpc error: code = NotFound desc = could not find container \"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c\": container with ID starting with fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c not found: ID does not exist" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.944224 4756 scope.go:117] "RemoveContainer" containerID="f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.944586 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533"} err="failed to get container status \"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533\": rpc error: code = NotFound desc = could not find container \"f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533\": container with ID starting with f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533 not found: ID does not exist" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.944603 4756 scope.go:117] "RemoveContainer" containerID="fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.944936 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c"} err="failed to get container status \"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c\": rpc error: code = NotFound desc = could not find container \"fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c\": container with ID starting with fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c not found: ID does not exist" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.962486 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmtgn\" (UniqueName: \"kubernetes.io/projected/cde1d7b5-d043-427a-86d8-471ebbd4fad4-kube-api-access-lmtgn\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.962522 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:13 crc kubenswrapper[4756]: I0203 09:30:13.962531 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cde1d7b5-d043-427a-86d8-471ebbd4fad4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.212185 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.232806 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.253553 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:14 crc kubenswrapper[4756]: E0203 09:30:14.254095 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerName="nova-metadata-log" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.254126 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerName="nova-metadata-log" Feb 03 09:30:14 crc kubenswrapper[4756]: E0203 09:30:14.254153 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerName="nova-metadata-metadata" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.254164 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerName="nova-metadata-metadata" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.254432 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerName="nova-metadata-log" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.254483 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" containerName="nova-metadata-metadata" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.256082 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.258704 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.262479 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.292403 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.372423 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.372759 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.372841 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgknq\" (UniqueName: \"kubernetes.io/projected/677d1c24-c53d-4416-8fc1-ebd4f4316996-kube-api-access-cgknq\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.372974 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/677d1c24-c53d-4416-8fc1-ebd4f4316996-logs\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.373262 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-config-data\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.474788 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-config-data\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.474848 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.474897 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.474927 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgknq\" (UniqueName: \"kubernetes.io/projected/677d1c24-c53d-4416-8fc1-ebd4f4316996-kube-api-access-cgknq\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.474977 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/677d1c24-c53d-4416-8fc1-ebd4f4316996-logs\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.475403 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/677d1c24-c53d-4416-8fc1-ebd4f4316996-logs\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.479183 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-config-data\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.491222 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.491732 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.494575 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgknq\" (UniqueName: \"kubernetes.io/projected/677d1c24-c53d-4416-8fc1-ebd4f4316996-kube-api-access-cgknq\") pod \"nova-metadata-0\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " pod="openstack/nova-metadata-0" Feb 03 09:30:14 crc kubenswrapper[4756]: I0203 09:30:14.573523 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:15 crc kubenswrapper[4756]: I0203 09:30:15.055979 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:15 crc kubenswrapper[4756]: I0203 09:30:15.627089 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cde1d7b5-d043-427a-86d8-471ebbd4fad4" path="/var/lib/kubelet/pods/cde1d7b5-d043-427a-86d8-471ebbd4fad4/volumes" Feb 03 09:30:15 crc kubenswrapper[4756]: I0203 09:30:15.891821 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"677d1c24-c53d-4416-8fc1-ebd4f4316996","Type":"ContainerStarted","Data":"4dab03173f02ccbd2e05645d59e37ef805c912a9dc556cfc852f83078dc364d5"} Feb 03 09:30:16 crc kubenswrapper[4756]: I0203 09:30:16.901428 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"677d1c24-c53d-4416-8fc1-ebd4f4316996","Type":"ContainerStarted","Data":"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a"} Feb 03 09:30:16 crc kubenswrapper[4756]: I0203 09:30:16.901717 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"677d1c24-c53d-4416-8fc1-ebd4f4316996","Type":"ContainerStarted","Data":"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909"} Feb 03 09:30:16 crc kubenswrapper[4756]: I0203 09:30:16.904427 4756 generic.go:334] "Generic (PLEG): container finished" podID="197062ac-2a3d-4b9b-a099-0d5843e4c58b" containerID="23531e769d889687f55e0157730d72e4aaae9107e14ac8e940390a190adeafbe" exitCode=0 Feb 03 09:30:16 crc kubenswrapper[4756]: I0203 09:30:16.904488 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7kx6w" event={"ID":"197062ac-2a3d-4b9b-a099-0d5843e4c58b","Type":"ContainerDied","Data":"23531e769d889687f55e0157730d72e4aaae9107e14ac8e940390a190adeafbe"} Feb 03 09:30:16 crc kubenswrapper[4756]: I0203 09:30:16.925347 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.925325438 podStartE2EDuration="2.925325438s" podCreationTimestamp="2026-02-03 09:30:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:16.921424975 +0000 UTC m=+1208.071892360" watchObservedRunningTime="2026-02-03 09:30:16.925325438 +0000 UTC m=+1208.075792813" Feb 03 09:30:17 crc kubenswrapper[4756]: I0203 09:30:17.921632 4756 generic.go:334] "Generic (PLEG): container finished" podID="5af01732-2a75-4a7d-af19-b097bc0ffd6b" containerID="81c525f541f93ae9180d083bfaf1b9f216fb3827bccee3cccc00d04cebf6a0ca" exitCode=0 Feb 03 09:30:17 crc kubenswrapper[4756]: I0203 09:30:17.922615 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" event={"ID":"5af01732-2a75-4a7d-af19-b097bc0ffd6b","Type":"ContainerDied","Data":"81c525f541f93ae9180d083bfaf1b9f216fb3827bccee3cccc00d04cebf6a0ca"} Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.032659 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.033773 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.066915 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.111635 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.294767 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.449062 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-scripts\") pod \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.450319 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-config-data\") pod \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.450374 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lvrg\" (UniqueName: \"kubernetes.io/projected/197062ac-2a3d-4b9b-a099-0d5843e4c58b-kube-api-access-5lvrg\") pod \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.450486 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-combined-ca-bundle\") pod \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\" (UID: \"197062ac-2a3d-4b9b-a099-0d5843e4c58b\") " Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.456698 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-scripts" (OuterVolumeSpecName: "scripts") pod "197062ac-2a3d-4b9b-a099-0d5843e4c58b" (UID: "197062ac-2a3d-4b9b-a099-0d5843e4c58b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.461649 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/197062ac-2a3d-4b9b-a099-0d5843e4c58b-kube-api-access-5lvrg" (OuterVolumeSpecName: "kube-api-access-5lvrg") pod "197062ac-2a3d-4b9b-a099-0d5843e4c58b" (UID: "197062ac-2a3d-4b9b-a099-0d5843e4c58b"). InnerVolumeSpecName "kube-api-access-5lvrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.481862 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-config-data" (OuterVolumeSpecName: "config-data") pod "197062ac-2a3d-4b9b-a099-0d5843e4c58b" (UID: "197062ac-2a3d-4b9b-a099-0d5843e4c58b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.488420 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.488469 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "197062ac-2a3d-4b9b-a099-0d5843e4c58b" (UID: "197062ac-2a3d-4b9b-a099-0d5843e4c58b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.552979 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.553012 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.553026 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lvrg\" (UniqueName: \"kubernetes.io/projected/197062ac-2a3d-4b9b-a099-0d5843e4c58b-kube-api-access-5lvrg\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.553042 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197062ac-2a3d-4b9b-a099-0d5843e4c58b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.563001 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-77mfv"] Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.563255 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" podUID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" containerName="dnsmasq-dns" containerID="cri-o://25f13ba113ff6857726f350d241462274b745ddd2f8b67087ade678372a050c4" gracePeriod=10 Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.939728 4756 generic.go:334] "Generic (PLEG): container finished" podID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" containerID="25f13ba113ff6857726f350d241462274b745ddd2f8b67087ade678372a050c4" exitCode=0 Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.939832 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" event={"ID":"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae","Type":"ContainerDied","Data":"25f13ba113ff6857726f350d241462274b745ddd2f8b67087ade678372a050c4"} Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.943420 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7kx6w" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.943365 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7kx6w" event={"ID":"197062ac-2a3d-4b9b-a099-0d5843e4c58b","Type":"ContainerDied","Data":"e4095f1c727b4bf683211b46a74ab8b90e3b8f6fa9937e1e235aaf000b43d415"} Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.960969 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4095f1c727b4bf683211b46a74ab8b90e3b8f6fa9937e1e235aaf000b43d415" Feb 03 09:30:18 crc kubenswrapper[4756]: I0203 09:30:18.977913 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.015006 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.062778 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-sb\") pod \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.062883 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-config\") pod \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.062924 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-swift-storage-0\") pod \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.062974 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-nb\") pod \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.063006 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-svc\") pod \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.063053 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6h8k\" (UniqueName: \"kubernetes.io/projected/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-kube-api-access-q6h8k\") pod \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\" (UID: \"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.080129 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-kube-api-access-q6h8k" (OuterVolumeSpecName: "kube-api-access-q6h8k") pod "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" (UID: "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae"). InnerVolumeSpecName "kube-api-access-q6h8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.121149 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.121534 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.141824 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-config" (OuterVolumeSpecName: "config") pod "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" (UID: "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.142280 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" (UID: "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.143647 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" (UID: "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.160862 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.167024 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" (UID: "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.172921 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.172955 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.172969 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.172980 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.172990 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6h8k\" (UniqueName: \"kubernetes.io/projected/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-kube-api-access-q6h8k\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.179226 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.183506 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerName="nova-metadata-log" containerID="cri-o://e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909" gracePeriod=30 Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.183687 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerName="nova-metadata-metadata" containerID="cri-o://d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a" gracePeriod=30 Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.229098 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" (UID: "59fdc32f-0ef0-40e8-b7c9-0eaf01017fae"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.274787 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.469597 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.560907 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.574939 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.580510 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.580895 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-combined-ca-bundle\") pod \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.580955 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-config-data\") pod \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.581036 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mp27\" (UniqueName: \"kubernetes.io/projected/5af01732-2a75-4a7d-af19-b097bc0ffd6b-kube-api-access-6mp27\") pod \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.581165 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-scripts\") pod \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\" (UID: \"5af01732-2a75-4a7d-af19-b097bc0ffd6b\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.586816 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-scripts" (OuterVolumeSpecName: "scripts") pod "5af01732-2a75-4a7d-af19-b097bc0ffd6b" (UID: "5af01732-2a75-4a7d-af19-b097bc0ffd6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.588948 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af01732-2a75-4a7d-af19-b097bc0ffd6b-kube-api-access-6mp27" (OuterVolumeSpecName: "kube-api-access-6mp27") pod "5af01732-2a75-4a7d-af19-b097bc0ffd6b" (UID: "5af01732-2a75-4a7d-af19-b097bc0ffd6b"). InnerVolumeSpecName "kube-api-access-6mp27". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.617559 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-config-data" (OuterVolumeSpecName: "config-data") pod "5af01732-2a75-4a7d-af19-b097bc0ffd6b" (UID: "5af01732-2a75-4a7d-af19-b097bc0ffd6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.631804 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5af01732-2a75-4a7d-af19-b097bc0ffd6b" (UID: "5af01732-2a75-4a7d-af19-b097bc0ffd6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.684692 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.684727 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.684738 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mp27\" (UniqueName: \"kubernetes.io/projected/5af01732-2a75-4a7d-af19-b097bc0ffd6b-kube-api-access-6mp27\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.684748 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af01732-2a75-4a7d-af19-b097bc0ffd6b-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.809152 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.887277 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/677d1c24-c53d-4416-8fc1-ebd4f4316996-logs\") pod \"677d1c24-c53d-4416-8fc1-ebd4f4316996\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.887360 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgknq\" (UniqueName: \"kubernetes.io/projected/677d1c24-c53d-4416-8fc1-ebd4f4316996-kube-api-access-cgknq\") pod \"677d1c24-c53d-4416-8fc1-ebd4f4316996\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.887472 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-config-data\") pod \"677d1c24-c53d-4416-8fc1-ebd4f4316996\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.887581 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-nova-metadata-tls-certs\") pod \"677d1c24-c53d-4416-8fc1-ebd4f4316996\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.887718 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-combined-ca-bundle\") pod \"677d1c24-c53d-4416-8fc1-ebd4f4316996\" (UID: \"677d1c24-c53d-4416-8fc1-ebd4f4316996\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.887942 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/677d1c24-c53d-4416-8fc1-ebd4f4316996-logs" (OuterVolumeSpecName: "logs") pod "677d1c24-c53d-4416-8fc1-ebd4f4316996" (UID: "677d1c24-c53d-4416-8fc1-ebd4f4316996"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.890312 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/677d1c24-c53d-4416-8fc1-ebd4f4316996-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.893509 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/677d1c24-c53d-4416-8fc1-ebd4f4316996-kube-api-access-cgknq" (OuterVolumeSpecName: "kube-api-access-cgknq") pod "677d1c24-c53d-4416-8fc1-ebd4f4316996" (UID: "677d1c24-c53d-4416-8fc1-ebd4f4316996"). InnerVolumeSpecName "kube-api-access-cgknq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.895554 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.916193 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "677d1c24-c53d-4416-8fc1-ebd4f4316996" (UID: "677d1c24-c53d-4416-8fc1-ebd4f4316996"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.939301 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-config-data" (OuterVolumeSpecName: "config-data") pod "677d1c24-c53d-4416-8fc1-ebd4f4316996" (UID: "677d1c24-c53d-4416-8fc1-ebd4f4316996"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.955379 4756 generic.go:334] "Generic (PLEG): container finished" podID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerID="d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a" exitCode=0 Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.955434 4756 generic.go:334] "Generic (PLEG): container finished" podID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerID="e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909" exitCode=143 Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.955480 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"677d1c24-c53d-4416-8fc1-ebd4f4316996","Type":"ContainerDied","Data":"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a"} Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.955522 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.955551 4756 scope.go:117] "RemoveContainer" containerID="d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.955537 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"677d1c24-c53d-4416-8fc1-ebd4f4316996","Type":"ContainerDied","Data":"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909"} Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.955725 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"677d1c24-c53d-4416-8fc1-ebd4f4316996","Type":"ContainerDied","Data":"4dab03173f02ccbd2e05645d59e37ef805c912a9dc556cfc852f83078dc364d5"} Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.959364 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" event={"ID":"5af01732-2a75-4a7d-af19-b097bc0ffd6b","Type":"ContainerDied","Data":"78e0aaaa531bb43cff4e6bbb00a1d4cb32f170a056cc7450cbbfe3de01d3d90c"} Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.959578 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78e0aaaa531bb43cff4e6bbb00a1d4cb32f170a056cc7450cbbfe3de01d3d90c" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.959775 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-4qgbq" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.965253 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.965292 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-77mfv" event={"ID":"59fdc32f-0ef0-40e8-b7c9-0eaf01017fae","Type":"ContainerDied","Data":"10d08e0a4eb1511a8f3d6db049319c60f27f248557a53da56e721af0b930e941"} Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.972175 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.973001 4756 generic.go:334] "Generic (PLEG): container finished" podID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerID="d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a" exitCode=137 Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.973059 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerDied","Data":"d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a"} Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.973085 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8cee259-a509-47ec-8abd-fb850e50cfb3","Type":"ContainerDied","Data":"bb3df286cc8930a7beef94f103e928bd8513ef5c5b1ec46212fb999fd49af0eb"} Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.973417 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-log" containerID="cri-o://d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf" gracePeriod=30 Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.973658 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-api" containerID="cri-o://052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc" gracePeriod=30 Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.982075 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "677d1c24-c53d-4416-8fc1-ebd4f4316996" (UID: "677d1c24-c53d-4416-8fc1-ebd4f4316996"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.991193 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-combined-ca-bundle\") pod \"b8cee259-a509-47ec-8abd-fb850e50cfb3\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.991254 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-scripts\") pod \"b8cee259-a509-47ec-8abd-fb850e50cfb3\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.991289 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-log-httpd\") pod \"b8cee259-a509-47ec-8abd-fb850e50cfb3\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.991325 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-sg-core-conf-yaml\") pod \"b8cee259-a509-47ec-8abd-fb850e50cfb3\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.991346 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-run-httpd\") pod \"b8cee259-a509-47ec-8abd-fb850e50cfb3\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.991514 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-config-data\") pod \"b8cee259-a509-47ec-8abd-fb850e50cfb3\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.991621 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j29j7\" (UniqueName: \"kubernetes.io/projected/b8cee259-a509-47ec-8abd-fb850e50cfb3-kube-api-access-j29j7\") pod \"b8cee259-a509-47ec-8abd-fb850e50cfb3\" (UID: \"b8cee259-a509-47ec-8abd-fb850e50cfb3\") " Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.992014 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.992031 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgknq\" (UniqueName: \"kubernetes.io/projected/677d1c24-c53d-4416-8fc1-ebd4f4316996-kube-api-access-cgknq\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.992045 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.992057 4756 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/677d1c24-c53d-4416-8fc1-ebd4f4316996-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.992599 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b8cee259-a509-47ec-8abd-fb850e50cfb3" (UID: "b8cee259-a509-47ec-8abd-fb850e50cfb3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:19 crc kubenswrapper[4756]: I0203 09:30:19.994114 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b8cee259-a509-47ec-8abd-fb850e50cfb3" (UID: "b8cee259-a509-47ec-8abd-fb850e50cfb3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.002690 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8cee259-a509-47ec-8abd-fb850e50cfb3-kube-api-access-j29j7" (OuterVolumeSpecName: "kube-api-access-j29j7") pod "b8cee259-a509-47ec-8abd-fb850e50cfb3" (UID: "b8cee259-a509-47ec-8abd-fb850e50cfb3"). InnerVolumeSpecName "kube-api-access-j29j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.002787 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-scripts" (OuterVolumeSpecName: "scripts") pod "b8cee259-a509-47ec-8abd-fb850e50cfb3" (UID: "b8cee259-a509-47ec-8abd-fb850e50cfb3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.015559 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-77mfv"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.018019 4756 scope.go:117] "RemoveContainer" containerID="e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.031233 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-77mfv"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.052651 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b8cee259-a509-47ec-8abd-fb850e50cfb3" (UID: "b8cee259-a509-47ec-8abd-fb850e50cfb3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.053911 4756 scope.go:117] "RemoveContainer" containerID="d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.054831 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a\": container with ID starting with d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a not found: ID does not exist" containerID="d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.054872 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a"} err="failed to get container status \"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a\": rpc error: code = NotFound desc = could not find container \"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a\": container with ID starting with d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a not found: ID does not exist" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.054898 4756 scope.go:117] "RemoveContainer" containerID="e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.055257 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909\": container with ID starting with e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909 not found: ID does not exist" containerID="e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.055287 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909"} err="failed to get container status \"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909\": rpc error: code = NotFound desc = could not find container \"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909\": container with ID starting with e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909 not found: ID does not exist" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.055310 4756 scope.go:117] "RemoveContainer" containerID="d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.055660 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a"} err="failed to get container status \"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a\": rpc error: code = NotFound desc = could not find container \"d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a\": container with ID starting with d7bae7b82a1f5b355e0931959a203e76a5884b78e60359e5406c44ad9629bb1a not found: ID does not exist" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.055682 4756 scope.go:117] "RemoveContainer" containerID="e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.056040 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909"} err="failed to get container status \"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909\": rpc error: code = NotFound desc = could not find container \"e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909\": container with ID starting with e1aabc9ccc6f37ab602ab5bfcc3975a34a03923258fd5794f92a93d108bd1909 not found: ID does not exist" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.056065 4756 scope.go:117] "RemoveContainer" containerID="25f13ba113ff6857726f350d241462274b745ddd2f8b67087ade678372a050c4" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.066024 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.066718 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af01732-2a75-4a7d-af19-b097bc0ffd6b" containerName="nova-cell1-conductor-db-sync" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.066840 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af01732-2a75-4a7d-af19-b097bc0ffd6b" containerName="nova-cell1-conductor-db-sync" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.066942 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="ceilometer-central-agent" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.067011 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="ceilometer-central-agent" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.067084 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerName="nova-metadata-metadata" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.067153 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerName="nova-metadata-metadata" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.067230 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" containerName="dnsmasq-dns" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.067299 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" containerName="dnsmasq-dns" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.067379 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" containerName="init" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.067495 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" containerName="init" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.067591 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerName="nova-metadata-log" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.067659 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerName="nova-metadata-log" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.067744 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="197062ac-2a3d-4b9b-a099-0d5843e4c58b" containerName="nova-manage" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.067817 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="197062ac-2a3d-4b9b-a099-0d5843e4c58b" containerName="nova-manage" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.067913 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="proxy-httpd" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.067987 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="proxy-httpd" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.068060 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="sg-core" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.068131 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="sg-core" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.068206 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="ceilometer-notification-agent" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.068277 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="ceilometer-notification-agent" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.068588 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af01732-2a75-4a7d-af19-b097bc0ffd6b" containerName="nova-cell1-conductor-db-sync" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.068692 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerName="nova-metadata-metadata" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.068770 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="ceilometer-notification-agent" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.068845 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="proxy-httpd" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.068933 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="ceilometer-central-agent" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.069017 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" containerName="sg-core" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.069096 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="197062ac-2a3d-4b9b-a099-0d5843e4c58b" containerName="nova-manage" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.069191 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" containerName="dnsmasq-dns" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.069274 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" containerName="nova-metadata-log" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.070112 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.072224 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.077217 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.093747 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j29j7\" (UniqueName: \"kubernetes.io/projected/b8cee259-a509-47ec-8abd-fb850e50cfb3-kube-api-access-j29j7\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.093779 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.093793 4756 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.093804 4756 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.093816 4756 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8cee259-a509-47ec-8abd-fb850e50cfb3-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.098597 4756 scope.go:117] "RemoveContainer" containerID="3ba8bbb10f06208e72abd6eb88537dc7a837d5ea2401f8b8305cfd5a7847149f" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.123556 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8cee259-a509-47ec-8abd-fb850e50cfb3" (UID: "b8cee259-a509-47ec-8abd-fb850e50cfb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.123610 4756 scope.go:117] "RemoveContainer" containerID="d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.153106 4756 scope.go:117] "RemoveContainer" containerID="3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.164888 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-config-data" (OuterVolumeSpecName: "config-data") pod "b8cee259-a509-47ec-8abd-fb850e50cfb3" (UID: "b8cee259-a509-47ec-8abd-fb850e50cfb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.174561 4756 scope.go:117] "RemoveContainer" containerID="8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.194887 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8gnf\" (UniqueName: \"kubernetes.io/projected/7978abb6-3f13-4b6e-8300-38a9d9cdd335-kube-api-access-s8gnf\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.194956 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7978abb6-3f13-4b6e-8300-38a9d9cdd335-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.194987 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7978abb6-3f13-4b6e-8300-38a9d9cdd335-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.195117 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.195130 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8cee259-a509-47ec-8abd-fb850e50cfb3-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.196379 4756 scope.go:117] "RemoveContainer" containerID="fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.235242 4756 scope.go:117] "RemoveContainer" containerID="d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.236123 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a\": container with ID starting with d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a not found: ID does not exist" containerID="d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.236578 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a"} err="failed to get container status \"d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a\": rpc error: code = NotFound desc = could not find container \"d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a\": container with ID starting with d73f9632be995aa2c766abf2f47e3423ac7963340d7a88d4f9e93263239e802a not found: ID does not exist" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.236617 4756 scope.go:117] "RemoveContainer" containerID="3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.237065 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe\": container with ID starting with 3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe not found: ID does not exist" containerID="3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.237094 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe"} err="failed to get container status \"3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe\": rpc error: code = NotFound desc = could not find container \"3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe\": container with ID starting with 3e0d8979b91693cd6dd8b5a94d138bfefc5ef4e13b2bad76eaaca62b1a58b9fe not found: ID does not exist" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.237115 4756 scope.go:117] "RemoveContainer" containerID="8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.237549 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966\": container with ID starting with 8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966 not found: ID does not exist" containerID="8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.237569 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966"} err="failed to get container status \"8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966\": rpc error: code = NotFound desc = could not find container \"8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966\": container with ID starting with 8eb7013b2c1d635e02ecfb1910ff236d8fdd980df5b9047a832c14c7b45e6966 not found: ID does not exist" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.237582 4756 scope.go:117] "RemoveContainer" containerID="fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06" Feb 03 09:30:20 crc kubenswrapper[4756]: E0203 09:30:20.238000 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06\": container with ID starting with fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06 not found: ID does not exist" containerID="fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.238019 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06"} err="failed to get container status \"fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06\": rpc error: code = NotFound desc = could not find container \"fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06\": container with ID starting with fb68d8823bbe0f68fe27addfb722562f424dcc335bcd17ab828b3f63bae7ed06 not found: ID does not exist" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.296268 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7978abb6-3f13-4b6e-8300-38a9d9cdd335-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.296410 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8gnf\" (UniqueName: \"kubernetes.io/projected/7978abb6-3f13-4b6e-8300-38a9d9cdd335-kube-api-access-s8gnf\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.296465 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7978abb6-3f13-4b6e-8300-38a9d9cdd335-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.300844 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7978abb6-3f13-4b6e-8300-38a9d9cdd335-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.306886 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7978abb6-3f13-4b6e-8300-38a9d9cdd335-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.314792 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8gnf\" (UniqueName: \"kubernetes.io/projected/7978abb6-3f13-4b6e-8300-38a9d9cdd335-kube-api-access-s8gnf\") pod \"nova-cell1-conductor-0\" (UID: \"7978abb6-3f13-4b6e-8300-38a9d9cdd335\") " pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.383146 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.392055 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.405978 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.427528 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.432169 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.449192 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.451255 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.464923 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.465309 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.465792 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.477158 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.496777 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.500402 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.500705 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.503415 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.602658 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.602905 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-config-data\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.602960 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-logs\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.602982 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-run-httpd\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.603000 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.603025 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfkmx\" (UniqueName: \"kubernetes.io/projected/9846d64b-f0ba-4d24-8601-e37c3a707a7b-kube-api-access-sfkmx\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.603048 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-log-httpd\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.603070 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-config-data\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.603100 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-scripts\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.603117 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.603162 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.603177 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc2bk\" (UniqueName: \"kubernetes.io/projected/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-kube-api-access-sc2bk\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705190 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-config-data\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705286 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-logs\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705316 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-run-httpd\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705337 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705367 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfkmx\" (UniqueName: \"kubernetes.io/projected/9846d64b-f0ba-4d24-8601-e37c3a707a7b-kube-api-access-sfkmx\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705400 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-log-httpd\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705425 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-config-data\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705508 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-scripts\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705532 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705584 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705601 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc2bk\" (UniqueName: \"kubernetes.io/projected/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-kube-api-access-sc2bk\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.705620 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.707310 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-log-httpd\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.707529 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-logs\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.707630 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-run-httpd\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.711901 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.712176 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.712236 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.712886 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.714268 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-config-data\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.717107 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-scripts\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.717666 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-config-data\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.725175 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc2bk\" (UniqueName: \"kubernetes.io/projected/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-kube-api-access-sc2bk\") pod \"nova-metadata-0\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.725370 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfkmx\" (UniqueName: \"kubernetes.io/projected/9846d64b-f0ba-4d24-8601-e37c3a707a7b-kube-api-access-sfkmx\") pod \"ceilometer-0\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.857703 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.865242 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:20 crc kubenswrapper[4756]: I0203 09:30:20.892319 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.004176 4756 generic.go:334] "Generic (PLEG): container finished" podID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerID="d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf" exitCode=143 Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.004369 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62679a7b-6a34-499e-b0c9-8a50b3afe493","Type":"ContainerDied","Data":"d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf"} Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.008848 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7978abb6-3f13-4b6e-8300-38a9d9cdd335","Type":"ContainerStarted","Data":"5d85ef67f0077d8f3bf7345159222fd0414c1da697fab22df14b6d5979d76db5"} Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.012393 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="83584248-5055-4851-8e03-e07f8eb0c306" containerName="nova-scheduler-scheduler" containerID="cri-o://2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f" gracePeriod=30 Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.344856 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.394657 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.625006 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59fdc32f-0ef0-40e8-b7c9-0eaf01017fae" path="/var/lib/kubelet/pods/59fdc32f-0ef0-40e8-b7c9-0eaf01017fae/volumes" Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.625777 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="677d1c24-c53d-4416-8fc1-ebd4f4316996" path="/var/lib/kubelet/pods/677d1c24-c53d-4416-8fc1-ebd4f4316996/volumes" Feb 03 09:30:21 crc kubenswrapper[4756]: I0203 09:30:21.626498 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8cee259-a509-47ec-8abd-fb850e50cfb3" path="/var/lib/kubelet/pods/b8cee259-a509-47ec-8abd-fb850e50cfb3/volumes" Feb 03 09:30:22 crc kubenswrapper[4756]: I0203 09:30:22.025648 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"7978abb6-3f13-4b6e-8300-38a9d9cdd335","Type":"ContainerStarted","Data":"9daea5a46a99418f2265e547f0bec4d6f66227338df4096436025dd8ddd41da3"} Feb 03 09:30:22 crc kubenswrapper[4756]: I0203 09:30:22.026805 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:22 crc kubenswrapper[4756]: I0203 09:30:22.031856 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerStarted","Data":"4efdd824ed3e60ec995eb7a0a5f875cc4ab42e36ea917e7ffe69d01b89c1222d"} Feb 03 09:30:22 crc kubenswrapper[4756]: I0203 09:30:22.035629 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee","Type":"ContainerStarted","Data":"39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0"} Feb 03 09:30:22 crc kubenswrapper[4756]: I0203 09:30:22.035707 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee","Type":"ContainerStarted","Data":"fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c"} Feb 03 09:30:22 crc kubenswrapper[4756]: I0203 09:30:22.035724 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee","Type":"ContainerStarted","Data":"4ead9934949f5387861a7c483a8bb7bdf0877eb0efc810b57fe3e4088cbba3b1"} Feb 03 09:30:22 crc kubenswrapper[4756]: I0203 09:30:22.055273 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.05525198 podStartE2EDuration="2.05525198s" podCreationTimestamp="2026-02-03 09:30:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:22.041951693 +0000 UTC m=+1213.192419068" watchObservedRunningTime="2026-02-03 09:30:22.05525198 +0000 UTC m=+1213.205719355" Feb 03 09:30:22 crc kubenswrapper[4756]: I0203 09:30:22.077176 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.077153935 podStartE2EDuration="2.077153935s" podCreationTimestamp="2026-02-03 09:30:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:22.062996862 +0000 UTC m=+1213.213464247" watchObservedRunningTime="2026-02-03 09:30:22.077153935 +0000 UTC m=+1213.227621320" Feb 03 09:30:23 crc kubenswrapper[4756]: I0203 09:30:23.055707 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerStarted","Data":"6dcdfd8928ef0320523191d3e8878652d87cd1f9c7b3edc8e920e946c582110b"} Feb 03 09:30:23 crc kubenswrapper[4756]: I0203 09:30:23.056120 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerStarted","Data":"95fd578561d1f538269e824863ee65b736b72455bd45a37f27b358850a627347"} Feb 03 09:30:23 crc kubenswrapper[4756]: E0203 09:30:23.076497 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 03 09:30:23 crc kubenswrapper[4756]: E0203 09:30:23.078213 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 03 09:30:23 crc kubenswrapper[4756]: E0203 09:30:23.079508 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 03 09:30:23 crc kubenswrapper[4756]: E0203 09:30:23.079550 4756 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="83584248-5055-4851-8e03-e07f8eb0c306" containerName="nova-scheduler-scheduler" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.068307 4756 generic.go:334] "Generic (PLEG): container finished" podID="83584248-5055-4851-8e03-e07f8eb0c306" containerID="2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f" exitCode=0 Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.068413 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83584248-5055-4851-8e03-e07f8eb0c306","Type":"ContainerDied","Data":"2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f"} Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.071706 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerStarted","Data":"c628844210f5ae4a4147223243b15a9d890d2320b389bf30bab36101a84e61cc"} Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.352056 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.481699 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-config-data\") pod \"83584248-5055-4851-8e03-e07f8eb0c306\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.482127 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksgtc\" (UniqueName: \"kubernetes.io/projected/83584248-5055-4851-8e03-e07f8eb0c306-kube-api-access-ksgtc\") pod \"83584248-5055-4851-8e03-e07f8eb0c306\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.482352 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-combined-ca-bundle\") pod \"83584248-5055-4851-8e03-e07f8eb0c306\" (UID: \"83584248-5055-4851-8e03-e07f8eb0c306\") " Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.490656 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83584248-5055-4851-8e03-e07f8eb0c306-kube-api-access-ksgtc" (OuterVolumeSpecName: "kube-api-access-ksgtc") pod "83584248-5055-4851-8e03-e07f8eb0c306" (UID: "83584248-5055-4851-8e03-e07f8eb0c306"). InnerVolumeSpecName "kube-api-access-ksgtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.514716 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-config-data" (OuterVolumeSpecName: "config-data") pod "83584248-5055-4851-8e03-e07f8eb0c306" (UID: "83584248-5055-4851-8e03-e07f8eb0c306"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.519685 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83584248-5055-4851-8e03-e07f8eb0c306" (UID: "83584248-5055-4851-8e03-e07f8eb0c306"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.584680 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.584711 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksgtc\" (UniqueName: \"kubernetes.io/projected/83584248-5055-4851-8e03-e07f8eb0c306-kube-api-access-ksgtc\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.584723 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83584248-5055-4851-8e03-e07f8eb0c306-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.877191 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.997474 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-combined-ca-bundle\") pod \"62679a7b-6a34-499e-b0c9-8a50b3afe493\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.997594 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62679a7b-6a34-499e-b0c9-8a50b3afe493-logs\") pod \"62679a7b-6a34-499e-b0c9-8a50b3afe493\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.997755 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7n9q\" (UniqueName: \"kubernetes.io/projected/62679a7b-6a34-499e-b0c9-8a50b3afe493-kube-api-access-j7n9q\") pod \"62679a7b-6a34-499e-b0c9-8a50b3afe493\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.997783 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-config-data\") pod \"62679a7b-6a34-499e-b0c9-8a50b3afe493\" (UID: \"62679a7b-6a34-499e-b0c9-8a50b3afe493\") " Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.998271 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62679a7b-6a34-499e-b0c9-8a50b3afe493-logs" (OuterVolumeSpecName: "logs") pod "62679a7b-6a34-499e-b0c9-8a50b3afe493" (UID: "62679a7b-6a34-499e-b0c9-8a50b3afe493"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:24 crc kubenswrapper[4756]: I0203 09:30:24.999048 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62679a7b-6a34-499e-b0c9-8a50b3afe493-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.003848 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62679a7b-6a34-499e-b0c9-8a50b3afe493-kube-api-access-j7n9q" (OuterVolumeSpecName: "kube-api-access-j7n9q") pod "62679a7b-6a34-499e-b0c9-8a50b3afe493" (UID: "62679a7b-6a34-499e-b0c9-8a50b3afe493"). InnerVolumeSpecName "kube-api-access-j7n9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.028056 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-config-data" (OuterVolumeSpecName: "config-data") pod "62679a7b-6a34-499e-b0c9-8a50b3afe493" (UID: "62679a7b-6a34-499e-b0c9-8a50b3afe493"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.028806 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62679a7b-6a34-499e-b0c9-8a50b3afe493" (UID: "62679a7b-6a34-499e-b0c9-8a50b3afe493"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.083469 4756 generic.go:334] "Generic (PLEG): container finished" podID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerID="052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc" exitCode=0 Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.083519 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62679a7b-6a34-499e-b0c9-8a50b3afe493","Type":"ContainerDied","Data":"052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc"} Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.083572 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62679a7b-6a34-499e-b0c9-8a50b3afe493","Type":"ContainerDied","Data":"fa77de810e7df0e67fcdbfb545de075bb1cf83ca87cce1f7198acef98627659d"} Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.083596 4756 scope.go:117] "RemoveContainer" containerID="052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.083799 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.088735 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83584248-5055-4851-8e03-e07f8eb0c306","Type":"ContainerDied","Data":"9e8e596f08f6597b6e3ba34cab1a3af5c3f8972cbf14c4a3298e986ad8624b7f"} Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.088795 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.100800 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.100862 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7n9q\" (UniqueName: \"kubernetes.io/projected/62679a7b-6a34-499e-b0c9-8a50b3afe493-kube-api-access-j7n9q\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.100875 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62679a7b-6a34-499e-b0c9-8a50b3afe493-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.130410 4756 scope.go:117] "RemoveContainer" containerID="d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.152410 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.167980 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.168042 4756 scope.go:117] "RemoveContainer" containerID="052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc" Feb 03 09:30:25 crc kubenswrapper[4756]: E0203 09:30:25.168425 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc\": container with ID starting with 052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc not found: ID does not exist" containerID="052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.168470 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc"} err="failed to get container status \"052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc\": rpc error: code = NotFound desc = could not find container \"052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc\": container with ID starting with 052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc not found: ID does not exist" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.168489 4756 scope.go:117] "RemoveContainer" containerID="d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf" Feb 03 09:30:25 crc kubenswrapper[4756]: E0203 09:30:25.168704 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf\": container with ID starting with d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf not found: ID does not exist" containerID="d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.168723 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf"} err="failed to get container status \"d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf\": rpc error: code = NotFound desc = could not find container \"d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf\": container with ID starting with d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf not found: ID does not exist" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.168737 4756 scope.go:117] "RemoveContainer" containerID="2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.184315 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.196292 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.217827 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:25 crc kubenswrapper[4756]: E0203 09:30:25.218220 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-log" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.218239 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-log" Feb 03 09:30:25 crc kubenswrapper[4756]: E0203 09:30:25.218259 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-api" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.218265 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-api" Feb 03 09:30:25 crc kubenswrapper[4756]: E0203 09:30:25.218287 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83584248-5055-4851-8e03-e07f8eb0c306" containerName="nova-scheduler-scheduler" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.218293 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="83584248-5055-4851-8e03-e07f8eb0c306" containerName="nova-scheduler-scheduler" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.218462 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-api" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.218476 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="83584248-5055-4851-8e03-e07f8eb0c306" containerName="nova-scheduler-scheduler" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.218499 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" containerName="nova-api-log" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.219144 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.221832 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.235357 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.239617 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.241573 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.243897 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.251660 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.303252 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg5n9\" (UniqueName: \"kubernetes.io/projected/cf5d8148-9082-4c45-bf23-8373a0e501f8-kube-api-access-kg5n9\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.303396 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5d8148-9082-4c45-bf23-8373a0e501f8-logs\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.303646 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-config-data\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.303742 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.303797 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.303887 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqs5s\" (UniqueName: \"kubernetes.io/projected/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-kube-api-access-kqs5s\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.303932 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-config-data\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.410538 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg5n9\" (UniqueName: \"kubernetes.io/projected/cf5d8148-9082-4c45-bf23-8373a0e501f8-kube-api-access-kg5n9\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.410603 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5d8148-9082-4c45-bf23-8373a0e501f8-logs\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.410662 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-config-data\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.410698 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.410732 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.410757 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqs5s\" (UniqueName: \"kubernetes.io/projected/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-kube-api-access-kqs5s\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.410777 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-config-data\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.411131 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5d8148-9082-4c45-bf23-8373a0e501f8-logs\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.417235 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.417303 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-config-data\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.422540 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.422622 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-config-data\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.431585 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg5n9\" (UniqueName: \"kubernetes.io/projected/cf5d8148-9082-4c45-bf23-8373a0e501f8-kube-api-access-kg5n9\") pod \"nova-api-0\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.433268 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqs5s\" (UniqueName: \"kubernetes.io/projected/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-kube-api-access-kqs5s\") pod \"nova-scheduler-0\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.544826 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.563072 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.637570 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62679a7b-6a34-499e-b0c9-8a50b3afe493" path="/var/lib/kubelet/pods/62679a7b-6a34-499e-b0c9-8a50b3afe493/volumes" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.638631 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83584248-5055-4851-8e03-e07f8eb0c306" path="/var/lib/kubelet/pods/83584248-5055-4851-8e03-e07f8eb0c306/volumes" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.858512 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 03 09:30:25 crc kubenswrapper[4756]: I0203 09:30:25.858582 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 03 09:30:26 crc kubenswrapper[4756]: I0203 09:30:26.032046 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:30:26 crc kubenswrapper[4756]: I0203 09:30:26.091582 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:26 crc kubenswrapper[4756]: I0203 09:30:26.110432 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerStarted","Data":"e8294815b9c028cf046376788bcb9685b6d587ee6dcf2d1eadaeb7a886c374f5"} Feb 03 09:30:26 crc kubenswrapper[4756]: I0203 09:30:26.110512 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 03 09:30:26 crc kubenswrapper[4756]: I0203 09:30:26.116531 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1d9da720-6d2e-43b4-bccf-f7c43aecea1c","Type":"ContainerStarted","Data":"88920b3dfde6d94c16352c1f17241d6bc6f29d9a51ffc19d8203846ea0cdcc79"} Feb 03 09:30:26 crc kubenswrapper[4756]: I0203 09:30:26.141883 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9222142089999998 podStartE2EDuration="6.141863409s" podCreationTimestamp="2026-02-03 09:30:20 +0000 UTC" firstStartedPulling="2026-02-03 09:30:21.404722793 +0000 UTC m=+1212.555190168" lastFinishedPulling="2026-02-03 09:30:25.624371993 +0000 UTC m=+1216.774839368" observedRunningTime="2026-02-03 09:30:26.135569632 +0000 UTC m=+1217.286037007" watchObservedRunningTime="2026-02-03 09:30:26.141863409 +0000 UTC m=+1217.292330774" Feb 03 09:30:27 crc kubenswrapper[4756]: I0203 09:30:27.132108 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1d9da720-6d2e-43b4-bccf-f7c43aecea1c","Type":"ContainerStarted","Data":"a3470984acb8918059791f041334c6cac1e3662e6314967dc00b29ba2944df9f"} Feb 03 09:30:27 crc kubenswrapper[4756]: I0203 09:30:27.137528 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf5d8148-9082-4c45-bf23-8373a0e501f8","Type":"ContainerStarted","Data":"46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e"} Feb 03 09:30:27 crc kubenswrapper[4756]: I0203 09:30:27.138203 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf5d8148-9082-4c45-bf23-8373a0e501f8","Type":"ContainerStarted","Data":"378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627"} Feb 03 09:30:27 crc kubenswrapper[4756]: I0203 09:30:27.138225 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf5d8148-9082-4c45-bf23-8373a0e501f8","Type":"ContainerStarted","Data":"d31a3d491b409ae2726dbff7c961e0ede1e47f5a9c20a278c3d700854389bcab"} Feb 03 09:30:27 crc kubenswrapper[4756]: I0203 09:30:27.157649 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.157626579 podStartE2EDuration="2.157626579s" podCreationTimestamp="2026-02-03 09:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:27.148731811 +0000 UTC m=+1218.299199186" watchObservedRunningTime="2026-02-03 09:30:27.157626579 +0000 UTC m=+1218.308093954" Feb 03 09:30:30 crc kubenswrapper[4756]: I0203 09:30:30.433845 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 03 09:30:30 crc kubenswrapper[4756]: I0203 09:30:30.451309 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=5.451284306 podStartE2EDuration="5.451284306s" podCreationTimestamp="2026-02-03 09:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:27.168427777 +0000 UTC m=+1218.318895172" watchObservedRunningTime="2026-02-03 09:30:30.451284306 +0000 UTC m=+1221.601751691" Feb 03 09:30:30 crc kubenswrapper[4756]: I0203 09:30:30.546561 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 03 09:30:30 crc kubenswrapper[4756]: I0203 09:30:30.859195 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 03 09:30:30 crc kubenswrapper[4756]: I0203 09:30:30.859243 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 03 09:30:31 crc kubenswrapper[4756]: I0203 09:30:31.871690 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 03 09:30:31 crc kubenswrapper[4756]: I0203 09:30:31.871728 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 03 09:30:35 crc kubenswrapper[4756]: I0203 09:30:35.546567 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 03 09:30:35 crc kubenswrapper[4756]: I0203 09:30:35.564780 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 03 09:30:35 crc kubenswrapper[4756]: I0203 09:30:35.564838 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 03 09:30:35 crc kubenswrapper[4756]: I0203 09:30:35.584498 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 03 09:30:36 crc kubenswrapper[4756]: I0203 09:30:36.248127 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 03 09:30:36 crc kubenswrapper[4756]: I0203 09:30:36.648702 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 03 09:30:36 crc kubenswrapper[4756]: I0203 09:30:36.648702 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 03 09:30:40 crc kubenswrapper[4756]: I0203 09:30:40.864271 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 03 09:30:40 crc kubenswrapper[4756]: I0203 09:30:40.865908 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 03 09:30:40 crc kubenswrapper[4756]: I0203 09:30:40.871691 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 03 09:30:41 crc kubenswrapper[4756]: I0203 09:30:41.272052 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 03 09:30:42 crc kubenswrapper[4756]: W0203 09:30:42.916210 4756 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod677d1c24_c53d_4416_8fc1_ebd4f4316996.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod677d1c24_c53d_4416_8fc1_ebd4f4316996.slice: no such file or directory Feb 03 09:30:42 crc kubenswrapper[4756]: W0203 09:30:42.917418 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83584248_5055_4851_8e03_e07f8eb0c306.slice/crio-2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f.scope WatchSource:0}: Error finding container 2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f: Status 404 returned error can't find the container with id 2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f Feb 03 09:30:42 crc kubenswrapper[4756]: W0203 09:30:42.917602 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62679a7b_6a34_499e_b0c9_8a50b3afe493.slice/crio-d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf.scope WatchSource:0}: Error finding container d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf: Status 404 returned error can't find the container with id d29e5f4e92eaa9a0e88bcd0e0fc86651d70320b18c395833dba50be01d69cdbf Feb 03 09:30:42 crc kubenswrapper[4756]: W0203 09:30:42.917790 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcde1d7b5_d043_427a_86d8_471ebbd4fad4.slice/crio-fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c.scope WatchSource:0}: Error finding container fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c: Status 404 returned error can't find the container with id fdbfd11db03e5d9c5ca4d2cb56964198828211a0aa22b94a268c93c37928f26c Feb 03 09:30:42 crc kubenswrapper[4756]: W0203 09:30:42.918369 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcde1d7b5_d043_427a_86d8_471ebbd4fad4.slice/crio-f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533.scope WatchSource:0}: Error finding container f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533: Status 404 returned error can't find the container with id f853a69a6d5210a480949ad61e0b27a41291c35ad34a9ad9ff975b1cb230c533 Feb 03 09:30:42 crc kubenswrapper[4756]: W0203 09:30:42.918625 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62679a7b_6a34_499e_b0c9_8a50b3afe493.slice/crio-052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc.scope WatchSource:0}: Error finding container 052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc: Status 404 returned error can't find the container with id 052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc Feb 03 09:30:43 crc kubenswrapper[4756]: E0203 09:30:43.148914 4756 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62679a7b_6a34_499e_b0c9_8a50b3afe493.slice/crio-conmon-052080ecdaacbb608890c2224ced7704d6f4ee2c4988c6820af660ed66291efc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83584248_5055_4851_8e03_e07f8eb0c306.slice/crio-9e8e596f08f6597b6e3ba34cab1a3af5c3f8972cbf14c4a3298e986ad8624b7f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f77a4df_b751_4e2e_a5bd_d22f4f453f81.slice/crio-2fe920db5b7ff2901fe8a303936aaa9ddb30bd04c5c80cf24c7208d76d56de96.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83584248_5055_4851_8e03_e07f8eb0c306.slice/crio-conmon-2d1f45f67956941311fa794423e88b76f3117e5f99a50ec850c778b929d6624f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83584248_5055_4851_8e03_e07f8eb0c306.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62679a7b_6a34_499e_b0c9_8a50b3afe493.slice/crio-fa77de810e7df0e67fcdbfb545de075bb1cf83ca87cce1f7198acef98627659d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f77a4df_b751_4e2e_a5bd_d22f4f453f81.slice/crio-conmon-2fe920db5b7ff2901fe8a303936aaa9ddb30bd04c5c80cf24c7208d76d56de96.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62679a7b_6a34_499e_b0c9_8a50b3afe493.slice\": RecentStats: unable to find data in memory cache]" Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.283908 4756 generic.go:334] "Generic (PLEG): container finished" podID="5f77a4df-b751-4e2e-a5bd-d22f4f453f81" containerID="2fe920db5b7ff2901fe8a303936aaa9ddb30bd04c5c80cf24c7208d76d56de96" exitCode=137 Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.283998 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5f77a4df-b751-4e2e-a5bd-d22f4f453f81","Type":"ContainerDied","Data":"2fe920db5b7ff2901fe8a303936aaa9ddb30bd04c5c80cf24c7208d76d56de96"} Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.284046 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5f77a4df-b751-4e2e-a5bd-d22f4f453f81","Type":"ContainerDied","Data":"30949ca8679459576e41152df1d0fc6d6ee30c53693c2182135e6fd50928824b"} Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.284058 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30949ca8679459576e41152df1d0fc6d6ee30c53693c2182135e6fd50928824b" Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.316035 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.429088 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-combined-ca-bundle\") pod \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.429556 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hzfq\" (UniqueName: \"kubernetes.io/projected/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-kube-api-access-4hzfq\") pod \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.429599 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-config-data\") pod \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\" (UID: \"5f77a4df-b751-4e2e-a5bd-d22f4f453f81\") " Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.436520 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-kube-api-access-4hzfq" (OuterVolumeSpecName: "kube-api-access-4hzfq") pod "5f77a4df-b751-4e2e-a5bd-d22f4f453f81" (UID: "5f77a4df-b751-4e2e-a5bd-d22f4f453f81"). InnerVolumeSpecName "kube-api-access-4hzfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.459589 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-config-data" (OuterVolumeSpecName: "config-data") pod "5f77a4df-b751-4e2e-a5bd-d22f4f453f81" (UID: "5f77a4df-b751-4e2e-a5bd-d22f4f453f81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.470292 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f77a4df-b751-4e2e-a5bd-d22f4f453f81" (UID: "5f77a4df-b751-4e2e-a5bd-d22f4f453f81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.531654 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.531691 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hzfq\" (UniqueName: \"kubernetes.io/projected/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-kube-api-access-4hzfq\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:43 crc kubenswrapper[4756]: I0203 09:30:43.531702 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f77a4df-b751-4e2e-a5bd-d22f4f453f81-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.291156 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.313017 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.321628 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.341518 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:44 crc kubenswrapper[4756]: E0203 09:30:44.341982 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f77a4df-b751-4e2e-a5bd-d22f4f453f81" containerName="nova-cell1-novncproxy-novncproxy" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.342001 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f77a4df-b751-4e2e-a5bd-d22f4f453f81" containerName="nova-cell1-novncproxy-novncproxy" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.342229 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f77a4df-b751-4e2e-a5bd-d22f4f453f81" containerName="nova-cell1-novncproxy-novncproxy" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.343916 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.347121 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.347206 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.355807 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.356360 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.451734 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.451848 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.451920 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.452168 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2q6m\" (UniqueName: \"kubernetes.io/projected/58183079-d9ac-4407-bec4-e93605aa884d-kube-api-access-p2q6m\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.452233 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.553950 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.554050 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.554123 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2q6m\" (UniqueName: \"kubernetes.io/projected/58183079-d9ac-4407-bec4-e93605aa884d-kube-api-access-p2q6m\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.554153 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.554206 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.559301 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.559416 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.559827 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.560057 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/58183079-d9ac-4407-bec4-e93605aa884d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.582381 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2q6m\" (UniqueName: \"kubernetes.io/projected/58183079-d9ac-4407-bec4-e93605aa884d-kube-api-access-p2q6m\") pod \"nova-cell1-novncproxy-0\" (UID: \"58183079-d9ac-4407-bec4-e93605aa884d\") " pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:44 crc kubenswrapper[4756]: I0203 09:30:44.666395 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:45 crc kubenswrapper[4756]: W0203 09:30:45.110845 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58183079_d9ac_4407_bec4_e93605aa884d.slice/crio-e36570fddc040b5c98f124fe7c6e821358f77cd040a666af120d4f72edd4782e WatchSource:0}: Error finding container e36570fddc040b5c98f124fe7c6e821358f77cd040a666af120d4f72edd4782e: Status 404 returned error can't find the container with id e36570fddc040b5c98f124fe7c6e821358f77cd040a666af120d4f72edd4782e Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.113102 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.300464 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"58183079-d9ac-4407-bec4-e93605aa884d","Type":"ContainerStarted","Data":"fd8d91f8ac2f1624f8c37224e5ae473942ec9a17201afbd80be29dea11f22a79"} Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.301798 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"58183079-d9ac-4407-bec4-e93605aa884d","Type":"ContainerStarted","Data":"e36570fddc040b5c98f124fe7c6e821358f77cd040a666af120d4f72edd4782e"} Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.571262 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.571814 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.577697 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.585171 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.621269 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.6212180379999999 podStartE2EDuration="1.621218038s" podCreationTimestamp="2026-02-03 09:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:45.316941365 +0000 UTC m=+1236.467408760" watchObservedRunningTime="2026-02-03 09:30:45.621218038 +0000 UTC m=+1236.771685403" Feb 03 09:30:45 crc kubenswrapper[4756]: I0203 09:30:45.630087 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f77a4df-b751-4e2e-a5bd-d22f4f453f81" path="/var/lib/kubelet/pods/5f77a4df-b751-4e2e-a5bd-d22f4f453f81/volumes" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.314294 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.321272 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.504505 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-qlqfx"] Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.512953 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.529042 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-qlqfx"] Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.602697 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-config\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.602754 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.602935 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6kzb\" (UniqueName: \"kubernetes.io/projected/97678a5a-ea1a-477f-8a7f-adff844186a4-kube-api-access-p6kzb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.603092 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.603171 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.603202 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.707595 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-config\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.707663 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.707707 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6kzb\" (UniqueName: \"kubernetes.io/projected/97678a5a-ea1a-477f-8a7f-adff844186a4-kube-api-access-p6kzb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.707762 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.707803 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.707863 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.709876 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-config\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.710382 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.711424 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.712002 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.712512 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.732346 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6kzb\" (UniqueName: \"kubernetes.io/projected/97678a5a-ea1a-477f-8a7f-adff844186a4-kube-api-access-p6kzb\") pod \"dnsmasq-dns-cd5cbd7b9-qlqfx\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:46 crc kubenswrapper[4756]: I0203 09:30:46.840348 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:47 crc kubenswrapper[4756]: I0203 09:30:47.337543 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-qlqfx"] Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.330911 4756 generic.go:334] "Generic (PLEG): container finished" podID="97678a5a-ea1a-477f-8a7f-adff844186a4" containerID="ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f" exitCode=0 Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.330971 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" event={"ID":"97678a5a-ea1a-477f-8a7f-adff844186a4","Type":"ContainerDied","Data":"ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f"} Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.331765 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" event={"ID":"97678a5a-ea1a-477f-8a7f-adff844186a4","Type":"ContainerStarted","Data":"ff9f147fa6cfe2691603a6204cb5028acb330b608fd84f21785568e3a8134e25"} Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.669673 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.670381 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="ceilometer-central-agent" containerID="cri-o://6dcdfd8928ef0320523191d3e8878652d87cd1f9c7b3edc8e920e946c582110b" gracePeriod=30 Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.670433 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="sg-core" containerID="cri-o://c628844210f5ae4a4147223243b15a9d890d2320b389bf30bab36101a84e61cc" gracePeriod=30 Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.670494 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="ceilometer-notification-agent" containerID="cri-o://95fd578561d1f538269e824863ee65b736b72455bd45a37f27b358850a627347" gracePeriod=30 Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.670529 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="proxy-httpd" containerID="cri-o://e8294815b9c028cf046376788bcb9685b6d587ee6dcf2d1eadaeb7a886c374f5" gracePeriod=30 Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.678817 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.196:3000/\": EOF" Feb 03 09:30:48 crc kubenswrapper[4756]: I0203 09:30:48.925854 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.367141 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" event={"ID":"97678a5a-ea1a-477f-8a7f-adff844186a4","Type":"ContainerStarted","Data":"cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5"} Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.367643 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.370632 4756 generic.go:334] "Generic (PLEG): container finished" podID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerID="e8294815b9c028cf046376788bcb9685b6d587ee6dcf2d1eadaeb7a886c374f5" exitCode=0 Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.370659 4756 generic.go:334] "Generic (PLEG): container finished" podID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerID="c628844210f5ae4a4147223243b15a9d890d2320b389bf30bab36101a84e61cc" exitCode=2 Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.370669 4756 generic.go:334] "Generic (PLEG): container finished" podID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerID="95fd578561d1f538269e824863ee65b736b72455bd45a37f27b358850a627347" exitCode=0 Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.370676 4756 generic.go:334] "Generic (PLEG): container finished" podID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerID="6dcdfd8928ef0320523191d3e8878652d87cd1f9c7b3edc8e920e946c582110b" exitCode=0 Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.370839 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-log" containerID="cri-o://378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627" gracePeriod=30 Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.371103 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerDied","Data":"e8294815b9c028cf046376788bcb9685b6d587ee6dcf2d1eadaeb7a886c374f5"} Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.371133 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerDied","Data":"c628844210f5ae4a4147223243b15a9d890d2320b389bf30bab36101a84e61cc"} Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.371144 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerDied","Data":"95fd578561d1f538269e824863ee65b736b72455bd45a37f27b358850a627347"} Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.371151 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerDied","Data":"6dcdfd8928ef0320523191d3e8878652d87cd1f9c7b3edc8e920e946c582110b"} Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.371195 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-api" containerID="cri-o://46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e" gracePeriod=30 Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.400576 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" podStartSLOduration=3.400559626 podStartE2EDuration="3.400559626s" podCreationTimestamp="2026-02-03 09:30:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:49.397691186 +0000 UTC m=+1240.548158561" watchObservedRunningTime="2026-02-03 09:30:49.400559626 +0000 UTC m=+1240.551027001" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.512105 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.576803 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-log-httpd\") pod \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.576903 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfkmx\" (UniqueName: \"kubernetes.io/projected/9846d64b-f0ba-4d24-8601-e37c3a707a7b-kube-api-access-sfkmx\") pod \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.576949 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-scripts\") pod \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.577064 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-config-data\") pod \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.577114 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-sg-core-conf-yaml\") pod \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.577199 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-run-httpd\") pod \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.577245 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-combined-ca-bundle\") pod \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\" (UID: \"9846d64b-f0ba-4d24-8601-e37c3a707a7b\") " Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.577296 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9846d64b-f0ba-4d24-8601-e37c3a707a7b" (UID: "9846d64b-f0ba-4d24-8601-e37c3a707a7b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.577683 4756 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.577844 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9846d64b-f0ba-4d24-8601-e37c3a707a7b" (UID: "9846d64b-f0ba-4d24-8601-e37c3a707a7b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.582902 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9846d64b-f0ba-4d24-8601-e37c3a707a7b-kube-api-access-sfkmx" (OuterVolumeSpecName: "kube-api-access-sfkmx") pod "9846d64b-f0ba-4d24-8601-e37c3a707a7b" (UID: "9846d64b-f0ba-4d24-8601-e37c3a707a7b"). InnerVolumeSpecName "kube-api-access-sfkmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.583574 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-scripts" (OuterVolumeSpecName: "scripts") pod "9846d64b-f0ba-4d24-8601-e37c3a707a7b" (UID: "9846d64b-f0ba-4d24-8601-e37c3a707a7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.614650 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9846d64b-f0ba-4d24-8601-e37c3a707a7b" (UID: "9846d64b-f0ba-4d24-8601-e37c3a707a7b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.660027 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9846d64b-f0ba-4d24-8601-e37c3a707a7b" (UID: "9846d64b-f0ba-4d24-8601-e37c3a707a7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.666861 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.679652 4756 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.679740 4756 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9846d64b-f0ba-4d24-8601-e37c3a707a7b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.680546 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.680562 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfkmx\" (UniqueName: \"kubernetes.io/projected/9846d64b-f0ba-4d24-8601-e37c3a707a7b-kube-api-access-sfkmx\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.680575 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.683792 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-config-data" (OuterVolumeSpecName: "config-data") pod "9846d64b-f0ba-4d24-8601-e37c3a707a7b" (UID: "9846d64b-f0ba-4d24-8601-e37c3a707a7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:49 crc kubenswrapper[4756]: I0203 09:30:49.782116 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9846d64b-f0ba-4d24-8601-e37c3a707a7b-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.380998 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9846d64b-f0ba-4d24-8601-e37c3a707a7b","Type":"ContainerDied","Data":"4efdd824ed3e60ec995eb7a0a5f875cc4ab42e36ea917e7ffe69d01b89c1222d"} Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.381055 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.381894 4756 scope.go:117] "RemoveContainer" containerID="e8294815b9c028cf046376788bcb9685b6d587ee6dcf2d1eadaeb7a886c374f5" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.384639 4756 generic.go:334] "Generic (PLEG): container finished" podID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerID="378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627" exitCode=143 Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.384968 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf5d8148-9082-4c45-bf23-8373a0e501f8","Type":"ContainerDied","Data":"378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627"} Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.418894 4756 scope.go:117] "RemoveContainer" containerID="c628844210f5ae4a4147223243b15a9d890d2320b389bf30bab36101a84e61cc" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.439680 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.444944 4756 scope.go:117] "RemoveContainer" containerID="95fd578561d1f538269e824863ee65b736b72455bd45a37f27b358850a627347" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.452470 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.469371 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:50 crc kubenswrapper[4756]: E0203 09:30:50.469835 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="proxy-httpd" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.469851 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="proxy-httpd" Feb 03 09:30:50 crc kubenswrapper[4756]: E0203 09:30:50.469870 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="ceilometer-notification-agent" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.469876 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="ceilometer-notification-agent" Feb 03 09:30:50 crc kubenswrapper[4756]: E0203 09:30:50.469888 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="sg-core" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.469894 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="sg-core" Feb 03 09:30:50 crc kubenswrapper[4756]: E0203 09:30:50.469905 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="ceilometer-central-agent" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.469910 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="ceilometer-central-agent" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.470081 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="sg-core" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.470092 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="proxy-httpd" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.470103 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="ceilometer-notification-agent" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.470114 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" containerName="ceilometer-central-agent" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.471727 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.475365 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.475411 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.477254 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.492650 4756 scope.go:117] "RemoveContainer" containerID="6dcdfd8928ef0320523191d3e8878652d87cd1f9c7b3edc8e920e946c582110b" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.556827 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:50 crc kubenswrapper[4756]: E0203 09:30:50.557688 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-wrzqx log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data kube-api-access-wrzqx log-httpd run-httpd scripts sg-core-conf-yaml]: context canceled" pod="openstack/ceilometer-0" podUID="92da459c-0963-4c81-8044-5a612f5bc95e" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.595068 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.595122 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-config-data\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.595162 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-run-httpd\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.595254 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.595285 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrzqx\" (UniqueName: \"kubernetes.io/projected/92da459c-0963-4c81-8044-5a612f5bc95e-kube-api-access-wrzqx\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.595342 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-scripts\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.595365 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-log-httpd\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.696638 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-run-httpd\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.696721 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.696771 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrzqx\" (UniqueName: \"kubernetes.io/projected/92da459c-0963-4c81-8044-5a612f5bc95e-kube-api-access-wrzqx\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.696821 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-scripts\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.696848 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-log-httpd\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.696973 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.696998 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-config-data\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.697072 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-run-httpd\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.698626 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-log-httpd\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.702179 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.702818 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.703028 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-scripts\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.711499 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-config-data\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:50 crc kubenswrapper[4756]: I0203 09:30:50.716521 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrzqx\" (UniqueName: \"kubernetes.io/projected/92da459c-0963-4c81-8044-5a612f5bc95e-kube-api-access-wrzqx\") pod \"ceilometer-0\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " pod="openstack/ceilometer-0" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.395161 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.407126 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.512822 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-combined-ca-bundle\") pod \"92da459c-0963-4c81-8044-5a612f5bc95e\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.512919 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-sg-core-conf-yaml\") pod \"92da459c-0963-4c81-8044-5a612f5bc95e\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.513006 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-scripts\") pod \"92da459c-0963-4c81-8044-5a612f5bc95e\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.513049 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-config-data\") pod \"92da459c-0963-4c81-8044-5a612f5bc95e\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.514015 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-log-httpd\") pod \"92da459c-0963-4c81-8044-5a612f5bc95e\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.514044 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrzqx\" (UniqueName: \"kubernetes.io/projected/92da459c-0963-4c81-8044-5a612f5bc95e-kube-api-access-wrzqx\") pod \"92da459c-0963-4c81-8044-5a612f5bc95e\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.514189 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-run-httpd\") pod \"92da459c-0963-4c81-8044-5a612f5bc95e\" (UID: \"92da459c-0963-4c81-8044-5a612f5bc95e\") " Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.514289 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "92da459c-0963-4c81-8044-5a612f5bc95e" (UID: "92da459c-0963-4c81-8044-5a612f5bc95e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.514657 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "92da459c-0963-4c81-8044-5a612f5bc95e" (UID: "92da459c-0963-4c81-8044-5a612f5bc95e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.515183 4756 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.515467 4756 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92da459c-0963-4c81-8044-5a612f5bc95e-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.518292 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-scripts" (OuterVolumeSpecName: "scripts") pod "92da459c-0963-4c81-8044-5a612f5bc95e" (UID: "92da459c-0963-4c81-8044-5a612f5bc95e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.518329 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92da459c-0963-4c81-8044-5a612f5bc95e-kube-api-access-wrzqx" (OuterVolumeSpecName: "kube-api-access-wrzqx") pod "92da459c-0963-4c81-8044-5a612f5bc95e" (UID: "92da459c-0963-4c81-8044-5a612f5bc95e"). InnerVolumeSpecName "kube-api-access-wrzqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.518627 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "92da459c-0963-4c81-8044-5a612f5bc95e" (UID: "92da459c-0963-4c81-8044-5a612f5bc95e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.519068 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-config-data" (OuterVolumeSpecName: "config-data") pod "92da459c-0963-4c81-8044-5a612f5bc95e" (UID: "92da459c-0963-4c81-8044-5a612f5bc95e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.523114 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92da459c-0963-4c81-8044-5a612f5bc95e" (UID: "92da459c-0963-4c81-8044-5a612f5bc95e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.616425 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.616472 4756 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.616481 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.616490 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92da459c-0963-4c81-8044-5a612f5bc95e-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.616498 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrzqx\" (UniqueName: \"kubernetes.io/projected/92da459c-0963-4c81-8044-5a612f5bc95e-kube-api-access-wrzqx\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:51 crc kubenswrapper[4756]: I0203 09:30:51.624010 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9846d64b-f0ba-4d24-8601-e37c3a707a7b" path="/var/lib/kubelet/pods/9846d64b-f0ba-4d24-8601-e37c3a707a7b/volumes" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.404688 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.535461 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.564489 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.572700 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.575032 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.577969 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.578051 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.587811 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.645380 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-log-httpd\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.645870 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-scripts\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.646100 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-config-data\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.646133 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqhpp\" (UniqueName: \"kubernetes.io/projected/3ca52f8b-dd1d-446b-a090-c1e4157176e1-kube-api-access-dqhpp\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.646215 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.646372 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.646526 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-run-httpd\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.747899 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-log-httpd\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.747998 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-scripts\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.748140 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-config-data\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.748166 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqhpp\" (UniqueName: \"kubernetes.io/projected/3ca52f8b-dd1d-446b-a090-c1e4157176e1-kube-api-access-dqhpp\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.748228 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.748288 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.748342 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-run-httpd\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.748875 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-log-httpd\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.748941 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-run-httpd\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.753864 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.753865 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.754540 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-config-data\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.762139 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-scripts\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.775237 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqhpp\" (UniqueName: \"kubernetes.io/projected/3ca52f8b-dd1d-446b-a090-c1e4157176e1-kube-api-access-dqhpp\") pod \"ceilometer-0\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.906521 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.951792 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-combined-ca-bundle\") pod \"cf5d8148-9082-4c45-bf23-8373a0e501f8\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.951924 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5d8148-9082-4c45-bf23-8373a0e501f8-logs\") pod \"cf5d8148-9082-4c45-bf23-8373a0e501f8\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.952011 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg5n9\" (UniqueName: \"kubernetes.io/projected/cf5d8148-9082-4c45-bf23-8373a0e501f8-kube-api-access-kg5n9\") pod \"cf5d8148-9082-4c45-bf23-8373a0e501f8\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.952046 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-config-data\") pod \"cf5d8148-9082-4c45-bf23-8373a0e501f8\" (UID: \"cf5d8148-9082-4c45-bf23-8373a0e501f8\") " Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.953641 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf5d8148-9082-4c45-bf23-8373a0e501f8-logs" (OuterVolumeSpecName: "logs") pod "cf5d8148-9082-4c45-bf23-8373a0e501f8" (UID: "cf5d8148-9082-4c45-bf23-8373a0e501f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.955960 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.970733 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf5d8148-9082-4c45-bf23-8373a0e501f8-kube-api-access-kg5n9" (OuterVolumeSpecName: "kube-api-access-kg5n9") pod "cf5d8148-9082-4c45-bf23-8373a0e501f8" (UID: "cf5d8148-9082-4c45-bf23-8373a0e501f8"). InnerVolumeSpecName "kube-api-access-kg5n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.978809 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf5d8148-9082-4c45-bf23-8373a0e501f8" (UID: "cf5d8148-9082-4c45-bf23-8373a0e501f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:52 crc kubenswrapper[4756]: I0203 09:30:52.985418 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-config-data" (OuterVolumeSpecName: "config-data") pod "cf5d8148-9082-4c45-bf23-8373a0e501f8" (UID: "cf5d8148-9082-4c45-bf23-8373a0e501f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.054195 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf5d8148-9082-4c45-bf23-8373a0e501f8-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.054244 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg5n9\" (UniqueName: \"kubernetes.io/projected/cf5d8148-9082-4c45-bf23-8373a0e501f8-kube-api-access-kg5n9\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.054262 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.054276 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf5d8148-9082-4c45-bf23-8373a0e501f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.418884 4756 generic.go:334] "Generic (PLEG): container finished" podID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerID="46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e" exitCode=0 Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.419305 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf5d8148-9082-4c45-bf23-8373a0e501f8","Type":"ContainerDied","Data":"46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e"} Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.419356 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf5d8148-9082-4c45-bf23-8373a0e501f8","Type":"ContainerDied","Data":"d31a3d491b409ae2726dbff7c961e0ede1e47f5a9c20a278c3d700854389bcab"} Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.419377 4756 scope.go:117] "RemoveContainer" containerID="46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.419471 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: W0203 09:30:53.435941 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ca52f8b_dd1d_446b_a090_c1e4157176e1.slice/crio-2f649f29095ded77d082ff7b4aa4de9ed77717ad5b9e9734577faf5c03227b8f WatchSource:0}: Error finding container 2f649f29095ded77d082ff7b4aa4de9ed77717ad5b9e9734577faf5c03227b8f: Status 404 returned error can't find the container with id 2f649f29095ded77d082ff7b4aa4de9ed77717ad5b9e9734577faf5c03227b8f Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.437160 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.442122 4756 scope.go:117] "RemoveContainer" containerID="378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.469025 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.480869 4756 scope.go:117] "RemoveContainer" containerID="46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e" Feb 03 09:30:53 crc kubenswrapper[4756]: E0203 09:30:53.481300 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e\": container with ID starting with 46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e not found: ID does not exist" containerID="46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.481342 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e"} err="failed to get container status \"46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e\": rpc error: code = NotFound desc = could not find container \"46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e\": container with ID starting with 46076139f64cd1e73a3ad1b8fbf3ed7903147e3c1e628c306df953fec4df8d0e not found: ID does not exist" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.481366 4756 scope.go:117] "RemoveContainer" containerID="378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627" Feb 03 09:30:53 crc kubenswrapper[4756]: E0203 09:30:53.481654 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627\": container with ID starting with 378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627 not found: ID does not exist" containerID="378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.481681 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627"} err="failed to get container status \"378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627\": rpc error: code = NotFound desc = could not find container \"378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627\": container with ID starting with 378718e1647646e5c8d10c80f737c3c7d9c62bfb536e9178541778082b44a627 not found: ID does not exist" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.483930 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.496478 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:53 crc kubenswrapper[4756]: E0203 09:30:53.497012 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-log" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.497036 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-log" Feb 03 09:30:53 crc kubenswrapper[4756]: E0203 09:30:53.497061 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-api" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.497071 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-api" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.497313 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-api" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.497341 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" containerName="nova-api-log" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.498560 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.501081 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.501133 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.501390 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.507439 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.561569 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-logs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.561612 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-config-data\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.561652 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-public-tls-certs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.561680 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92pd7\" (UniqueName: \"kubernetes.io/projected/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-kube-api-access-92pd7\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.561719 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.561750 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.626299 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92da459c-0963-4c81-8044-5a612f5bc95e" path="/var/lib/kubelet/pods/92da459c-0963-4c81-8044-5a612f5bc95e/volumes" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.626820 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf5d8148-9082-4c45-bf23-8373a0e501f8" path="/var/lib/kubelet/pods/cf5d8148-9082-4c45-bf23-8373a0e501f8/volumes" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.663259 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-logs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.663324 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-config-data\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.663351 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-public-tls-certs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.663373 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92pd7\" (UniqueName: \"kubernetes.io/projected/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-kube-api-access-92pd7\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.663403 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.663431 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.663816 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-logs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.668154 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-public-tls-certs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.668639 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.670169 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-config-data\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.677615 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.680286 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92pd7\" (UniqueName: \"kubernetes.io/projected/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-kube-api-access-92pd7\") pod \"nova-api-0\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " pod="openstack/nova-api-0" Feb 03 09:30:53 crc kubenswrapper[4756]: I0203 09:30:53.826396 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:30:54 crc kubenswrapper[4756]: W0203 09:30:54.343592 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc1d71f3_5c53_48bc_9d6c_8a5f0e10d3a8.slice/crio-3774085bc1b89b16e8ec5c0e1ef704f926d80a1b375eb6e5bd8ce85fd73f1ce9 WatchSource:0}: Error finding container 3774085bc1b89b16e8ec5c0e1ef704f926d80a1b375eb6e5bd8ce85fd73f1ce9: Status 404 returned error can't find the container with id 3774085bc1b89b16e8ec5c0e1ef704f926d80a1b375eb6e5bd8ce85fd73f1ce9 Feb 03 09:30:54 crc kubenswrapper[4756]: I0203 09:30:54.344712 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:30:54 crc kubenswrapper[4756]: I0203 09:30:54.428483 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8","Type":"ContainerStarted","Data":"3774085bc1b89b16e8ec5c0e1ef704f926d80a1b375eb6e5bd8ce85fd73f1ce9"} Feb 03 09:30:54 crc kubenswrapper[4756]: I0203 09:30:54.430291 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerStarted","Data":"bbcbd081925ae74379917495609fb1d18e3156dbcee7f2b3c972d0a9d38cc618"} Feb 03 09:30:54 crc kubenswrapper[4756]: I0203 09:30:54.430340 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerStarted","Data":"2f649f29095ded77d082ff7b4aa4de9ed77717ad5b9e9734577faf5c03227b8f"} Feb 03 09:30:54 crc kubenswrapper[4756]: I0203 09:30:54.666680 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:54 crc kubenswrapper[4756]: I0203 09:30:54.684848 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.443783 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8","Type":"ContainerStarted","Data":"936d4ddfbad84d71090688ba2d7ea6e1d5e4bfc558f243446d08a9d6339686a1"} Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.444061 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8","Type":"ContainerStarted","Data":"478585637a4c698a0a8007b3105b7f26fdca2a838eb61dce5934385134095b6b"} Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.446147 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerStarted","Data":"01edc87ebba40d6af61ecd48708d7c0f128ea4ffc8b8e29e0467a010a577e84e"} Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.478972 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.486477 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.486432464 podStartE2EDuration="2.486432464s" podCreationTimestamp="2026-02-03 09:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:55.466270164 +0000 UTC m=+1246.616737579" watchObservedRunningTime="2026-02-03 09:30:55.486432464 +0000 UTC m=+1246.636899839" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.671783 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mw4b6"] Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.673435 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.676556 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.676756 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.717719 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mw4b6"] Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.719091 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-config-data\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.719835 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hcsh\" (UniqueName: \"kubernetes.io/projected/50d3d3a0-68e5-449d-94fb-b295fd0d073e-kube-api-access-4hcsh\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.719865 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-scripts\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.719889 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.821641 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hcsh\" (UniqueName: \"kubernetes.io/projected/50d3d3a0-68e5-449d-94fb-b295fd0d073e-kube-api-access-4hcsh\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.821694 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-scripts\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.821720 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.821807 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-config-data\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.837741 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-scripts\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.837744 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.838917 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-config-data\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:55 crc kubenswrapper[4756]: I0203 09:30:55.841146 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hcsh\" (UniqueName: \"kubernetes.io/projected/50d3d3a0-68e5-449d-94fb-b295fd0d073e-kube-api-access-4hcsh\") pod \"nova-cell1-cell-mapping-mw4b6\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:56 crc kubenswrapper[4756]: I0203 09:30:56.050664 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:30:56 crc kubenswrapper[4756]: I0203 09:30:56.467506 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerStarted","Data":"13ba6bda4e41fd693ffb839cd47a69928c65ae3146960eca54c63fe08033b092"} Feb 03 09:30:56 crc kubenswrapper[4756]: I0203 09:30:56.534024 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mw4b6"] Feb 03 09:30:56 crc kubenswrapper[4756]: W0203 09:30:56.543142 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50d3d3a0_68e5_449d_94fb_b295fd0d073e.slice/crio-5344a9b80c2e4c2a3831c6dd34f152ad844c8ed5a5e63058f6facfbcc5fec821 WatchSource:0}: Error finding container 5344a9b80c2e4c2a3831c6dd34f152ad844c8ed5a5e63058f6facfbcc5fec821: Status 404 returned error can't find the container with id 5344a9b80c2e4c2a3831c6dd34f152ad844c8ed5a5e63058f6facfbcc5fec821 Feb 03 09:30:56 crc kubenswrapper[4756]: I0203 09:30:56.841599 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:30:56 crc kubenswrapper[4756]: I0203 09:30:56.916337 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-vp5rl"] Feb 03 09:30:56 crc kubenswrapper[4756]: I0203 09:30:56.916666 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" podUID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" containerName="dnsmasq-dns" containerID="cri-o://b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30" gracePeriod=10 Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.427781 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.452111 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-sb\") pod \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.452204 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-svc\") pod \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.452241 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-nb\") pod \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.452269 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-config\") pod \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.452305 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-swift-storage-0\") pod \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.452593 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz2j2\" (UniqueName: \"kubernetes.io/projected/f34a237b-60ac-4350-a1c3-6dc43b603fd7-kube-api-access-nz2j2\") pod \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\" (UID: \"f34a237b-60ac-4350-a1c3-6dc43b603fd7\") " Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.457752 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f34a237b-60ac-4350-a1c3-6dc43b603fd7-kube-api-access-nz2j2" (OuterVolumeSpecName: "kube-api-access-nz2j2") pod "f34a237b-60ac-4350-a1c3-6dc43b603fd7" (UID: "f34a237b-60ac-4350-a1c3-6dc43b603fd7"). InnerVolumeSpecName "kube-api-access-nz2j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.487320 4756 generic.go:334] "Generic (PLEG): container finished" podID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" containerID="b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30" exitCode=0 Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.487389 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" event={"ID":"f34a237b-60ac-4350-a1c3-6dc43b603fd7","Type":"ContainerDied","Data":"b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30"} Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.487423 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" event={"ID":"f34a237b-60ac-4350-a1c3-6dc43b603fd7","Type":"ContainerDied","Data":"02f338939498c8fd9041b5a032a6005ac3f42c1a01ce39c09d95dddba525890e"} Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.487516 4756 scope.go:117] "RemoveContainer" containerID="b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.487685 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-vp5rl" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.492292 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mw4b6" event={"ID":"50d3d3a0-68e5-449d-94fb-b295fd0d073e","Type":"ContainerStarted","Data":"593f53ccf44138863664fff0bd20206942d11fe64a3a5e7f228ce19684693c23"} Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.492390 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mw4b6" event={"ID":"50d3d3a0-68e5-449d-94fb-b295fd0d073e","Type":"ContainerStarted","Data":"5344a9b80c2e4c2a3831c6dd34f152ad844c8ed5a5e63058f6facfbcc5fec821"} Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.522489 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mw4b6" podStartSLOduration=2.522468748 podStartE2EDuration="2.522468748s" podCreationTimestamp="2026-02-03 09:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:30:57.512119425 +0000 UTC m=+1248.662586810" watchObservedRunningTime="2026-02-03 09:30:57.522468748 +0000 UTC m=+1248.672936123" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.539108 4756 scope.go:117] "RemoveContainer" containerID="fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.544533 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f34a237b-60ac-4350-a1c3-6dc43b603fd7" (UID: "f34a237b-60ac-4350-a1c3-6dc43b603fd7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.554465 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz2j2\" (UniqueName: \"kubernetes.io/projected/f34a237b-60ac-4350-a1c3-6dc43b603fd7-kube-api-access-nz2j2\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.554491 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.558130 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f34a237b-60ac-4350-a1c3-6dc43b603fd7" (UID: "f34a237b-60ac-4350-a1c3-6dc43b603fd7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.583731 4756 scope.go:117] "RemoveContainer" containerID="b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30" Feb 03 09:30:57 crc kubenswrapper[4756]: E0203 09:30:57.588214 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30\": container with ID starting with b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30 not found: ID does not exist" containerID="b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.588283 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30"} err="failed to get container status \"b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30\": rpc error: code = NotFound desc = could not find container \"b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30\": container with ID starting with b1a39be639c85ee966489897e4f286c27861392587e1a5632e0ffecc9d4c4e30 not found: ID does not exist" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.588306 4756 scope.go:117] "RemoveContainer" containerID="fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb" Feb 03 09:30:57 crc kubenswrapper[4756]: E0203 09:30:57.588744 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb\": container with ID starting with fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb not found: ID does not exist" containerID="fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.588764 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb"} err="failed to get container status \"fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb\": rpc error: code = NotFound desc = could not find container \"fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb\": container with ID starting with fd02971c915da8c98f054a30d4776ed96cd3d4754cb516531644b437779536eb not found: ID does not exist" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.592564 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f34a237b-60ac-4350-a1c3-6dc43b603fd7" (UID: "f34a237b-60ac-4350-a1c3-6dc43b603fd7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.602511 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-config" (OuterVolumeSpecName: "config") pod "f34a237b-60ac-4350-a1c3-6dc43b603fd7" (UID: "f34a237b-60ac-4350-a1c3-6dc43b603fd7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.620579 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f34a237b-60ac-4350-a1c3-6dc43b603fd7" (UID: "f34a237b-60ac-4350-a1c3-6dc43b603fd7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.656844 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.656881 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.656894 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.656904 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f34a237b-60ac-4350-a1c3-6dc43b603fd7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.811553 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-vp5rl"] Feb 03 09:30:57 crc kubenswrapper[4756]: I0203 09:30:57.820137 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-vp5rl"] Feb 03 09:30:59 crc kubenswrapper[4756]: I0203 09:30:59.516779 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerStarted","Data":"3c157f626593654ad5af135ca36c79030fd04fed353f1a93b0fac268ec847efb"} Feb 03 09:30:59 crc kubenswrapper[4756]: I0203 09:30:59.518232 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 03 09:30:59 crc kubenswrapper[4756]: I0203 09:30:59.548424 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.642647062 podStartE2EDuration="7.548403197s" podCreationTimestamp="2026-02-03 09:30:52 +0000 UTC" firstStartedPulling="2026-02-03 09:30:53.442247066 +0000 UTC m=+1244.592714441" lastFinishedPulling="2026-02-03 09:30:58.348003201 +0000 UTC m=+1249.498470576" observedRunningTime="2026-02-03 09:30:59.542173203 +0000 UTC m=+1250.692640588" watchObservedRunningTime="2026-02-03 09:30:59.548403197 +0000 UTC m=+1250.698870572" Feb 03 09:30:59 crc kubenswrapper[4756]: I0203 09:30:59.626919 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" path="/var/lib/kubelet/pods/f34a237b-60ac-4350-a1c3-6dc43b603fd7/volumes" Feb 03 09:31:02 crc kubenswrapper[4756]: I0203 09:31:02.619631 4756 generic.go:334] "Generic (PLEG): container finished" podID="50d3d3a0-68e5-449d-94fb-b295fd0d073e" containerID="593f53ccf44138863664fff0bd20206942d11fe64a3a5e7f228ce19684693c23" exitCode=0 Feb 03 09:31:02 crc kubenswrapper[4756]: I0203 09:31:02.619670 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mw4b6" event={"ID":"50d3d3a0-68e5-449d-94fb-b295fd0d073e","Type":"ContainerDied","Data":"593f53ccf44138863664fff0bd20206942d11fe64a3a5e7f228ce19684693c23"} Feb 03 09:31:03 crc kubenswrapper[4756]: I0203 09:31:03.827408 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 03 09:31:03 crc kubenswrapper[4756]: I0203 09:31:03.827475 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.006269 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.181858 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-scripts\") pod \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.181926 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-config-data\") pod \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.182003 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-combined-ca-bundle\") pod \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.182130 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hcsh\" (UniqueName: \"kubernetes.io/projected/50d3d3a0-68e5-449d-94fb-b295fd0d073e-kube-api-access-4hcsh\") pod \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\" (UID: \"50d3d3a0-68e5-449d-94fb-b295fd0d073e\") " Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.189118 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-scripts" (OuterVolumeSpecName: "scripts") pod "50d3d3a0-68e5-449d-94fb-b295fd0d073e" (UID: "50d3d3a0-68e5-449d-94fb-b295fd0d073e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.189566 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50d3d3a0-68e5-449d-94fb-b295fd0d073e-kube-api-access-4hcsh" (OuterVolumeSpecName: "kube-api-access-4hcsh") pod "50d3d3a0-68e5-449d-94fb-b295fd0d073e" (UID: "50d3d3a0-68e5-449d-94fb-b295fd0d073e"). InnerVolumeSpecName "kube-api-access-4hcsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.210897 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-config-data" (OuterVolumeSpecName: "config-data") pod "50d3d3a0-68e5-449d-94fb-b295fd0d073e" (UID: "50d3d3a0-68e5-449d-94fb-b295fd0d073e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.227887 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50d3d3a0-68e5-449d-94fb-b295fd0d073e" (UID: "50d3d3a0-68e5-449d-94fb-b295fd0d073e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.283882 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hcsh\" (UniqueName: \"kubernetes.io/projected/50d3d3a0-68e5-449d-94fb-b295fd0d073e-kube-api-access-4hcsh\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.283916 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.283925 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.283937 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50d3d3a0-68e5-449d-94fb-b295fd0d073e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.636822 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mw4b6" event={"ID":"50d3d3a0-68e5-449d-94fb-b295fd0d073e","Type":"ContainerDied","Data":"5344a9b80c2e4c2a3831c6dd34f152ad844c8ed5a5e63058f6facfbcc5fec821"} Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.636861 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5344a9b80c2e4c2a3831c6dd34f152ad844c8ed5a5e63058f6facfbcc5fec821" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.636878 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mw4b6" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.822089 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.822352 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-log" containerID="cri-o://478585637a4c698a0a8007b3105b7f26fdca2a838eb61dce5934385134095b6b" gracePeriod=30 Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.822533 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-api" containerID="cri-o://936d4ddfbad84d71090688ba2d7ea6e1d5e4bfc558f243446d08a9d6339686a1" gracePeriod=30 Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.835863 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.835975 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.841549 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.841824 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1d9da720-6d2e-43b4-bccf-f7c43aecea1c" containerName="nova-scheduler-scheduler" containerID="cri-o://a3470984acb8918059791f041334c6cac1e3662e6314967dc00b29ba2944df9f" gracePeriod=30 Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.864998 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.865331 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-log" containerID="cri-o://fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c" gracePeriod=30 Feb 03 09:31:04 crc kubenswrapper[4756]: I0203 09:31:04.865558 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-metadata" containerID="cri-o://39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0" gracePeriod=30 Feb 03 09:31:05 crc kubenswrapper[4756]: E0203 09:31:05.549091 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a3470984acb8918059791f041334c6cac1e3662e6314967dc00b29ba2944df9f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 03 09:31:05 crc kubenswrapper[4756]: E0203 09:31:05.550677 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a3470984acb8918059791f041334c6cac1e3662e6314967dc00b29ba2944df9f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 03 09:31:05 crc kubenswrapper[4756]: E0203 09:31:05.555706 4756 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a3470984acb8918059791f041334c6cac1e3662e6314967dc00b29ba2944df9f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 03 09:31:05 crc kubenswrapper[4756]: E0203 09:31:05.555767 4756 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1d9da720-6d2e-43b4-bccf-f7c43aecea1c" containerName="nova-scheduler-scheduler" Feb 03 09:31:05 crc kubenswrapper[4756]: I0203 09:31:05.647903 4756 generic.go:334] "Generic (PLEG): container finished" podID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerID="fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c" exitCode=143 Feb 03 09:31:05 crc kubenswrapper[4756]: I0203 09:31:05.647956 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee","Type":"ContainerDied","Data":"fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c"} Feb 03 09:31:05 crc kubenswrapper[4756]: I0203 09:31:05.650095 4756 generic.go:334] "Generic (PLEG): container finished" podID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerID="478585637a4c698a0a8007b3105b7f26fdca2a838eb61dce5934385134095b6b" exitCode=143 Feb 03 09:31:05 crc kubenswrapper[4756]: I0203 09:31:05.650131 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8","Type":"ContainerDied","Data":"478585637a4c698a0a8007b3105b7f26fdca2a838eb61dce5934385134095b6b"} Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.044350 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:60042->10.217.0.195:8775: read: connection reset by peer" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.044430 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:60040->10.217.0.195:8775: read: connection reset by peer" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.484197 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.661874 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc2bk\" (UniqueName: \"kubernetes.io/projected/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-kube-api-access-sc2bk\") pod \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.662095 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-nova-metadata-tls-certs\") pod \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.662176 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-logs\") pod \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.662218 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-combined-ca-bundle\") pod \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.662244 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-config-data\") pod \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\" (UID: \"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee\") " Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.662707 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-logs" (OuterVolumeSpecName: "logs") pod "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" (UID: "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.663427 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.670341 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-kube-api-access-sc2bk" (OuterVolumeSpecName: "kube-api-access-sc2bk") pod "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" (UID: "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee"). InnerVolumeSpecName "kube-api-access-sc2bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.682688 4756 generic.go:334] "Generic (PLEG): container finished" podID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerID="39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0" exitCode=0 Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.682747 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.682772 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee","Type":"ContainerDied","Data":"39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0"} Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.683129 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1a47160-72ed-4dc1-9f6d-9d475b1b0aee","Type":"ContainerDied","Data":"4ead9934949f5387861a7c483a8bb7bdf0877eb0efc810b57fe3e4088cbba3b1"} Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.683147 4756 scope.go:117] "RemoveContainer" containerID="39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.707979 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-config-data" (OuterVolumeSpecName: "config-data") pod "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" (UID: "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.712618 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" (UID: "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.720493 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" (UID: "a1a47160-72ed-4dc1-9f6d-9d475b1b0aee"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.765368 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.765403 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.765413 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc2bk\" (UniqueName: \"kubernetes.io/projected/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-kube-api-access-sc2bk\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.765424 4756 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.774130 4756 scope.go:117] "RemoveContainer" containerID="fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.792949 4756 scope.go:117] "RemoveContainer" containerID="39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0" Feb 03 09:31:08 crc kubenswrapper[4756]: E0203 09:31:08.793439 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0\": container with ID starting with 39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0 not found: ID does not exist" containerID="39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.793510 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0"} err="failed to get container status \"39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0\": rpc error: code = NotFound desc = could not find container \"39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0\": container with ID starting with 39e90b43665ecb447fba1760a4297262a059b1869536e0fce84985cf56b729e0 not found: ID does not exist" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.793538 4756 scope.go:117] "RemoveContainer" containerID="fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c" Feb 03 09:31:08 crc kubenswrapper[4756]: E0203 09:31:08.793859 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c\": container with ID starting with fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c not found: ID does not exist" containerID="fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c" Feb 03 09:31:08 crc kubenswrapper[4756]: I0203 09:31:08.793890 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c"} err="failed to get container status \"fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c\": rpc error: code = NotFound desc = could not find container \"fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c\": container with ID starting with fb5d3a3d0e5a85a78d171ef6de26d38fa4cdcdc218c9a286d7183bb9a72f023c not found: ID does not exist" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.020903 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.028570 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.043252 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:31:09 crc kubenswrapper[4756]: E0203 09:31:09.046852 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-metadata" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.046888 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-metadata" Feb 03 09:31:09 crc kubenswrapper[4756]: E0203 09:31:09.046911 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-log" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.046918 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-log" Feb 03 09:31:09 crc kubenswrapper[4756]: E0203 09:31:09.046958 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" containerName="dnsmasq-dns" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.046967 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" containerName="dnsmasq-dns" Feb 03 09:31:09 crc kubenswrapper[4756]: E0203 09:31:09.046992 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50d3d3a0-68e5-449d-94fb-b295fd0d073e" containerName="nova-manage" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.046998 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="50d3d3a0-68e5-449d-94fb-b295fd0d073e" containerName="nova-manage" Feb 03 09:31:09 crc kubenswrapper[4756]: E0203 09:31:09.047011 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" containerName="init" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.047018 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" containerName="init" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.047233 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f34a237b-60ac-4350-a1c3-6dc43b603fd7" containerName="dnsmasq-dns" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.047243 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-log" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.047254 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="50d3d3a0-68e5-449d-94fb-b295fd0d073e" containerName="nova-manage" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.047263 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" containerName="nova-metadata-metadata" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.048438 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.050743 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.051477 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.073788 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.174412 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.174499 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776f1390-6de5-4054-bff2-58621c198147-logs\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.174679 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.174714 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-config-data\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.174737 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbr9m\" (UniqueName: \"kubernetes.io/projected/776f1390-6de5-4054-bff2-58621c198147-kube-api-access-nbr9m\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.280499 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.280583 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-config-data\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.280611 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbr9m\" (UniqueName: \"kubernetes.io/projected/776f1390-6de5-4054-bff2-58621c198147-kube-api-access-nbr9m\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.280680 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.280708 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776f1390-6de5-4054-bff2-58621c198147-logs\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.281304 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776f1390-6de5-4054-bff2-58621c198147-logs\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.286264 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.286313 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.286319 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776f1390-6de5-4054-bff2-58621c198147-config-data\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.297077 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbr9m\" (UniqueName: \"kubernetes.io/projected/776f1390-6de5-4054-bff2-58621c198147-kube-api-access-nbr9m\") pod \"nova-metadata-0\" (UID: \"776f1390-6de5-4054-bff2-58621c198147\") " pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.367733 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.637471 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1a47160-72ed-4dc1-9f6d-9d475b1b0aee" path="/var/lib/kubelet/pods/a1a47160-72ed-4dc1-9f6d-9d475b1b0aee/volumes" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.710457 4756 generic.go:334] "Generic (PLEG): container finished" podID="1d9da720-6d2e-43b4-bccf-f7c43aecea1c" containerID="a3470984acb8918059791f041334c6cac1e3662e6314967dc00b29ba2944df9f" exitCode=0 Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.710521 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1d9da720-6d2e-43b4-bccf-f7c43aecea1c","Type":"ContainerDied","Data":"a3470984acb8918059791f041334c6cac1e3662e6314967dc00b29ba2944df9f"} Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.797219 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.813515 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.992151 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqs5s\" (UniqueName: \"kubernetes.io/projected/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-kube-api-access-kqs5s\") pod \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.992284 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-combined-ca-bundle\") pod \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.992353 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-config-data\") pod \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\" (UID: \"1d9da720-6d2e-43b4-bccf-f7c43aecea1c\") " Feb 03 09:31:09 crc kubenswrapper[4756]: I0203 09:31:09.996813 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-kube-api-access-kqs5s" (OuterVolumeSpecName: "kube-api-access-kqs5s") pod "1d9da720-6d2e-43b4-bccf-f7c43aecea1c" (UID: "1d9da720-6d2e-43b4-bccf-f7c43aecea1c"). InnerVolumeSpecName "kube-api-access-kqs5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.018938 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-config-data" (OuterVolumeSpecName: "config-data") pod "1d9da720-6d2e-43b4-bccf-f7c43aecea1c" (UID: "1d9da720-6d2e-43b4-bccf-f7c43aecea1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.026318 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d9da720-6d2e-43b4-bccf-f7c43aecea1c" (UID: "1d9da720-6d2e-43b4-bccf-f7c43aecea1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.094964 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqs5s\" (UniqueName: \"kubernetes.io/projected/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-kube-api-access-kqs5s\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.095001 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.095010 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d9da720-6d2e-43b4-bccf-f7c43aecea1c-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.722263 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.722251 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1d9da720-6d2e-43b4-bccf-f7c43aecea1c","Type":"ContainerDied","Data":"88920b3dfde6d94c16352c1f17241d6bc6f29d9a51ffc19d8203846ea0cdcc79"} Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.723275 4756 scope.go:117] "RemoveContainer" containerID="a3470984acb8918059791f041334c6cac1e3662e6314967dc00b29ba2944df9f" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.726244 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776f1390-6de5-4054-bff2-58621c198147","Type":"ContainerStarted","Data":"8a7481b45188c74d17ec44f87870f1517e5552e7a64b3f42c3a0825884f1b727"} Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.726281 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776f1390-6de5-4054-bff2-58621c198147","Type":"ContainerStarted","Data":"731914f66d8b6281f54475c6c0dcf8ca1a5c4230733b1d94d839de46f9384944"} Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.726295 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"776f1390-6de5-4054-bff2-58621c198147","Type":"ContainerStarted","Data":"fc99835bba1e8aa902c78f7a23a4da56d480dfdfc3bc756fb8184e16e4900b3d"} Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.731131 4756 generic.go:334] "Generic (PLEG): container finished" podID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerID="936d4ddfbad84d71090688ba2d7ea6e1d5e4bfc558f243446d08a9d6339686a1" exitCode=0 Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.731183 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8","Type":"ContainerDied","Data":"936d4ddfbad84d71090688ba2d7ea6e1d5e4bfc558f243446d08a9d6339686a1"} Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.731206 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8","Type":"ContainerDied","Data":"3774085bc1b89b16e8ec5c0e1ef704f926d80a1b375eb6e5bd8ce85fd73f1ce9"} Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.731216 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3774085bc1b89b16e8ec5c0e1ef704f926d80a1b375eb6e5bd8ce85fd73f1ce9" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.744952 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.760971 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.760953797 podStartE2EDuration="1.760953797s" podCreationTimestamp="2026-02-03 09:31:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:31:10.749274253 +0000 UTC m=+1261.899741618" watchObservedRunningTime="2026-02-03 09:31:10.760953797 +0000 UTC m=+1261.911421172" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.793259 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.808847 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.820986 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:31:10 crc kubenswrapper[4756]: E0203 09:31:10.821429 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-api" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.821464 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-api" Feb 03 09:31:10 crc kubenswrapper[4756]: E0203 09:31:10.821476 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-log" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.821484 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-log" Feb 03 09:31:10 crc kubenswrapper[4756]: E0203 09:31:10.821498 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9da720-6d2e-43b4-bccf-f7c43aecea1c" containerName="nova-scheduler-scheduler" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.821505 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9da720-6d2e-43b4-bccf-f7c43aecea1c" containerName="nova-scheduler-scheduler" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.821695 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-api" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.821718 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" containerName="nova-api-log" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.821729 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d9da720-6d2e-43b4-bccf-f7c43aecea1c" containerName="nova-scheduler-scheduler" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.822349 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.825566 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.828126 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.911363 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92pd7\" (UniqueName: \"kubernetes.io/projected/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-kube-api-access-92pd7\") pod \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.911463 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-combined-ca-bundle\") pod \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.911517 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-logs\") pod \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.911587 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-internal-tls-certs\") pod \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.911718 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-config-data\") pod \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.911756 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-public-tls-certs\") pod \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\" (UID: \"bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8\") " Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.912043 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-logs" (OuterVolumeSpecName: "logs") pod "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" (UID: "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.913461 4756 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-logs\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.916467 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-kube-api-access-92pd7" (OuterVolumeSpecName: "kube-api-access-92pd7") pod "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" (UID: "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8"). InnerVolumeSpecName "kube-api-access-92pd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.940595 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-config-data" (OuterVolumeSpecName: "config-data") pod "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" (UID: "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.941431 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" (UID: "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.960919 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" (UID: "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:10 crc kubenswrapper[4756]: I0203 09:31:10.966038 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" (UID: "bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.014520 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fead6bc0-9b80-4ab0-938c-19fad62f8178-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.014627 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fead6bc0-9b80-4ab0-938c-19fad62f8178-config-data\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.014660 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhcp4\" (UniqueName: \"kubernetes.io/projected/fead6bc0-9b80-4ab0-938c-19fad62f8178-kube-api-access-fhcp4\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.014738 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92pd7\" (UniqueName: \"kubernetes.io/projected/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-kube-api-access-92pd7\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.014749 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.014758 4756 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.014766 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.014774 4756 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.118812 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fead6bc0-9b80-4ab0-938c-19fad62f8178-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.119917 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fead6bc0-9b80-4ab0-938c-19fad62f8178-config-data\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.120056 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhcp4\" (UniqueName: \"kubernetes.io/projected/fead6bc0-9b80-4ab0-938c-19fad62f8178-kube-api-access-fhcp4\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.124148 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fead6bc0-9b80-4ab0-938c-19fad62f8178-config-data\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.124206 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fead6bc0-9b80-4ab0-938c-19fad62f8178-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.141506 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhcp4\" (UniqueName: \"kubernetes.io/projected/fead6bc0-9b80-4ab0-938c-19fad62f8178-kube-api-access-fhcp4\") pod \"nova-scheduler-0\" (UID: \"fead6bc0-9b80-4ab0-938c-19fad62f8178\") " pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.437932 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.647685 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d9da720-6d2e-43b4-bccf-f7c43aecea1c" path="/var/lib/kubelet/pods/1d9da720-6d2e-43b4-bccf-f7c43aecea1c/volumes" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.740922 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.764963 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.783749 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.792159 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.793942 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.795989 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.796228 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.803727 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.806028 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.907219 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.945575 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g5t8\" (UniqueName: \"kubernetes.io/projected/ca27353e-8cea-4f67-a90b-9776e510e82e-kube-api-access-6g5t8\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.945635 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.945667 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.945805 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-public-tls-certs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.946063 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca27353e-8cea-4f67-a90b-9776e510e82e-logs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:11 crc kubenswrapper[4756]: I0203 09:31:11.946266 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-config-data\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.050678 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-config-data\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.050840 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g5t8\" (UniqueName: \"kubernetes.io/projected/ca27353e-8cea-4f67-a90b-9776e510e82e-kube-api-access-6g5t8\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.050872 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.050898 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.050971 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-public-tls-certs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.051013 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca27353e-8cea-4f67-a90b-9776e510e82e-logs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.051967 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca27353e-8cea-4f67-a90b-9776e510e82e-logs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.056163 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.057144 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-config-data\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.062438 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.066162 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca27353e-8cea-4f67-a90b-9776e510e82e-public-tls-certs\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.074323 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g5t8\" (UniqueName: \"kubernetes.io/projected/ca27353e-8cea-4f67-a90b-9776e510e82e-kube-api-access-6g5t8\") pod \"nova-api-0\" (UID: \"ca27353e-8cea-4f67-a90b-9776e510e82e\") " pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.110902 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.535480 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 03 09:31:12 crc kubenswrapper[4756]: W0203 09:31:12.536242 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca27353e_8cea_4f67_a90b_9776e510e82e.slice/crio-9b28e065c378054250d7c4c1bea7726bcc5f5c498327560d78c05d1665a0ff7c WatchSource:0}: Error finding container 9b28e065c378054250d7c4c1bea7726bcc5f5c498327560d78c05d1665a0ff7c: Status 404 returned error can't find the container with id 9b28e065c378054250d7c4c1bea7726bcc5f5c498327560d78c05d1665a0ff7c Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.757075 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca27353e-8cea-4f67-a90b-9776e510e82e","Type":"ContainerStarted","Data":"1236f7c71333ac26ec22d433ddabe6a25b9d706e3ce2ba2c21066c23ce54d25d"} Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.757118 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca27353e-8cea-4f67-a90b-9776e510e82e","Type":"ContainerStarted","Data":"9b28e065c378054250d7c4c1bea7726bcc5f5c498327560d78c05d1665a0ff7c"} Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.758389 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fead6bc0-9b80-4ab0-938c-19fad62f8178","Type":"ContainerStarted","Data":"238897deb32f8555627731c513bf8424eb41e14e42e6374086eb55d50101ea94"} Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.758426 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fead6bc0-9b80-4ab0-938c-19fad62f8178","Type":"ContainerStarted","Data":"b46d814091f432c6a47a52524d263fe472ea11f69a55ef7c0df0c330b8a291ce"} Feb 03 09:31:12 crc kubenswrapper[4756]: I0203 09:31:12.780933 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7809122200000003 podStartE2EDuration="2.78091222s" podCreationTimestamp="2026-02-03 09:31:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:31:12.773361255 +0000 UTC m=+1263.923828640" watchObservedRunningTime="2026-02-03 09:31:12.78091222 +0000 UTC m=+1263.931379595" Feb 03 09:31:13 crc kubenswrapper[4756]: I0203 09:31:13.566637 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:31:13 crc kubenswrapper[4756]: I0203 09:31:13.567003 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:31:13 crc kubenswrapper[4756]: I0203 09:31:13.625122 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8" path="/var/lib/kubelet/pods/bc1d71f3-5c53-48bc-9d6c-8a5f0e10d3a8/volumes" Feb 03 09:31:13 crc kubenswrapper[4756]: I0203 09:31:13.774638 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca27353e-8cea-4f67-a90b-9776e510e82e","Type":"ContainerStarted","Data":"c37b276b2554726253b295d63a37f052d91ed8c0b36be96769fcb4f53176ea96"} Feb 03 09:31:13 crc kubenswrapper[4756]: I0203 09:31:13.796735 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.796690979 podStartE2EDuration="2.796690979s" podCreationTimestamp="2026-02-03 09:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:31:13.792071034 +0000 UTC m=+1264.942538429" watchObservedRunningTime="2026-02-03 09:31:13.796690979 +0000 UTC m=+1264.947158354" Feb 03 09:31:14 crc kubenswrapper[4756]: I0203 09:31:14.368939 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 03 09:31:14 crc kubenswrapper[4756]: I0203 09:31:14.369239 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 03 09:31:16 crc kubenswrapper[4756]: I0203 09:31:16.438465 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 03 09:31:19 crc kubenswrapper[4756]: I0203 09:31:19.368824 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 03 09:31:19 crc kubenswrapper[4756]: I0203 09:31:19.369202 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 03 09:31:20 crc kubenswrapper[4756]: I0203 09:31:20.382703 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="776f1390-6de5-4054-bff2-58621c198147" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 03 09:31:20 crc kubenswrapper[4756]: I0203 09:31:20.382703 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="776f1390-6de5-4054-bff2-58621c198147" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 03 09:31:21 crc kubenswrapper[4756]: I0203 09:31:21.438295 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 03 09:31:21 crc kubenswrapper[4756]: I0203 09:31:21.465502 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 03 09:31:21 crc kubenswrapper[4756]: I0203 09:31:21.882009 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 03 09:31:22 crc kubenswrapper[4756]: I0203 09:31:22.111647 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 03 09:31:22 crc kubenswrapper[4756]: I0203 09:31:22.111932 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 03 09:31:22 crc kubenswrapper[4756]: I0203 09:31:22.967865 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 03 09:31:23 crc kubenswrapper[4756]: I0203 09:31:23.126685 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ca27353e-8cea-4f67-a90b-9776e510e82e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 03 09:31:23 crc kubenswrapper[4756]: I0203 09:31:23.126697 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ca27353e-8cea-4f67-a90b-9776e510e82e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 03 09:31:26 crc kubenswrapper[4756]: I0203 09:31:26.602022 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:31:26 crc kubenswrapper[4756]: I0203 09:31:26.604630 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="773ac9a0-544a-4c28-a100-59f10ec789c7" containerName="kube-state-metrics" containerID="cri-o://1791a77fd71211fb49839eac63107ebf2ba24862a135dcb5b1f85769c0591e6d" gracePeriod=30 Feb 03 09:31:26 crc kubenswrapper[4756]: I0203 09:31:26.888815 4756 generic.go:334] "Generic (PLEG): container finished" podID="773ac9a0-544a-4c28-a100-59f10ec789c7" containerID="1791a77fd71211fb49839eac63107ebf2ba24862a135dcb5b1f85769c0591e6d" exitCode=2 Feb 03 09:31:26 crc kubenswrapper[4756]: I0203 09:31:26.889094 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"773ac9a0-544a-4c28-a100-59f10ec789c7","Type":"ContainerDied","Data":"1791a77fd71211fb49839eac63107ebf2ba24862a135dcb5b1f85769c0591e6d"} Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.065598 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.141908 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbwqv\" (UniqueName: \"kubernetes.io/projected/773ac9a0-544a-4c28-a100-59f10ec789c7-kube-api-access-qbwqv\") pod \"773ac9a0-544a-4c28-a100-59f10ec789c7\" (UID: \"773ac9a0-544a-4c28-a100-59f10ec789c7\") " Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.153738 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/773ac9a0-544a-4c28-a100-59f10ec789c7-kube-api-access-qbwqv" (OuterVolumeSpecName: "kube-api-access-qbwqv") pod "773ac9a0-544a-4c28-a100-59f10ec789c7" (UID: "773ac9a0-544a-4c28-a100-59f10ec789c7"). InnerVolumeSpecName "kube-api-access-qbwqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.246193 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbwqv\" (UniqueName: \"kubernetes.io/projected/773ac9a0-544a-4c28-a100-59f10ec789c7-kube-api-access-qbwqv\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.901907 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"773ac9a0-544a-4c28-a100-59f10ec789c7","Type":"ContainerDied","Data":"14a37fb11a14511a909627cd2e65b9d03bc79e6d4ea2e09bdf33e92e931cf813"} Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.902000 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.902259 4756 scope.go:117] "RemoveContainer" containerID="1791a77fd71211fb49839eac63107ebf2ba24862a135dcb5b1f85769c0591e6d" Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.952512 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:31:27 crc kubenswrapper[4756]: I0203 09:31:27.970749 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.002297 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:31:28 crc kubenswrapper[4756]: E0203 09:31:28.002878 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773ac9a0-544a-4c28-a100-59f10ec789c7" containerName="kube-state-metrics" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.002900 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="773ac9a0-544a-4c28-a100-59f10ec789c7" containerName="kube-state-metrics" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.003191 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="773ac9a0-544a-4c28-a100-59f10ec789c7" containerName="kube-state-metrics" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.004266 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.007368 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.007635 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.040692 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.161036 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.161112 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.161183 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.161269 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kzwd\" (UniqueName: \"kubernetes.io/projected/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-api-access-8kzwd\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.263349 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.263420 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kzwd\" (UniqueName: \"kubernetes.io/projected/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-api-access-8kzwd\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.263552 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.263634 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.268676 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.276129 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.280311 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.282934 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kzwd\" (UniqueName: \"kubernetes.io/projected/dc0c1050-2088-4a2d-98bc-fb14078d6418-kube-api-access-8kzwd\") pod \"kube-state-metrics-0\" (UID: \"dc0c1050-2088-4a2d-98bc-fb14078d6418\") " pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.320055 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.520186 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.522389 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="ceilometer-central-agent" containerID="cri-o://bbcbd081925ae74379917495609fb1d18e3156dbcee7f2b3c972d0a9d38cc618" gracePeriod=30 Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.522428 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="proxy-httpd" containerID="cri-o://3c157f626593654ad5af135ca36c79030fd04fed353f1a93b0fac268ec847efb" gracePeriod=30 Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.522627 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="sg-core" containerID="cri-o://13ba6bda4e41fd693ffb839cd47a69928c65ae3146960eca54c63fe08033b092" gracePeriod=30 Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.522648 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="ceilometer-notification-agent" containerID="cri-o://01edc87ebba40d6af61ecd48708d7c0f128ea4ffc8b8e29e0467a010a577e84e" gracePeriod=30 Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.795369 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.912336 4756 generic.go:334] "Generic (PLEG): container finished" podID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerID="3c157f626593654ad5af135ca36c79030fd04fed353f1a93b0fac268ec847efb" exitCode=0 Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.912373 4756 generic.go:334] "Generic (PLEG): container finished" podID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerID="13ba6bda4e41fd693ffb839cd47a69928c65ae3146960eca54c63fe08033b092" exitCode=2 Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.912421 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerDied","Data":"3c157f626593654ad5af135ca36c79030fd04fed353f1a93b0fac268ec847efb"} Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.912505 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerDied","Data":"13ba6bda4e41fd693ffb839cd47a69928c65ae3146960eca54c63fe08033b092"} Feb 03 09:31:28 crc kubenswrapper[4756]: I0203 09:31:28.913916 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"dc0c1050-2088-4a2d-98bc-fb14078d6418","Type":"ContainerStarted","Data":"442b5c55b61bae48b6efb355d29f3c046acbfefdfa754da9a7f371117a1fd854"} Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.374016 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.374678 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.380310 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.623706 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="773ac9a0-544a-4c28-a100-59f10ec789c7" path="/var/lib/kubelet/pods/773ac9a0-544a-4c28-a100-59f10ec789c7/volumes" Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.927029 4756 generic.go:334] "Generic (PLEG): container finished" podID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerID="bbcbd081925ae74379917495609fb1d18e3156dbcee7f2b3c972d0a9d38cc618" exitCode=0 Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.927116 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerDied","Data":"bbcbd081925ae74379917495609fb1d18e3156dbcee7f2b3c972d0a9d38cc618"} Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.930127 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"dc0c1050-2088-4a2d-98bc-fb14078d6418","Type":"ContainerStarted","Data":"795ecf31f41923e0ae1dac0973023e4a77fbe2a59d08f43ee50a98a5ec322228"} Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.930409 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.934645 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 03 09:31:29 crc kubenswrapper[4756]: I0203 09:31:29.974110 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.542097932 podStartE2EDuration="2.974091397s" podCreationTimestamp="2026-02-03 09:31:27 +0000 UTC" firstStartedPulling="2026-02-03 09:31:28.807913789 +0000 UTC m=+1279.958381164" lastFinishedPulling="2026-02-03 09:31:29.239907244 +0000 UTC m=+1280.390374629" observedRunningTime="2026-02-03 09:31:29.951865604 +0000 UTC m=+1281.102332979" watchObservedRunningTime="2026-02-03 09:31:29.974091397 +0000 UTC m=+1281.124558762" Feb 03 09:31:31 crc kubenswrapper[4756]: I0203 09:31:31.950023 4756 generic.go:334] "Generic (PLEG): container finished" podID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerID="01edc87ebba40d6af61ecd48708d7c0f128ea4ffc8b8e29e0467a010a577e84e" exitCode=0 Feb 03 09:31:31 crc kubenswrapper[4756]: I0203 09:31:31.950092 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerDied","Data":"01edc87ebba40d6af61ecd48708d7c0f128ea4ffc8b8e29e0467a010a577e84e"} Feb 03 09:31:31 crc kubenswrapper[4756]: I0203 09:31:31.950594 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ca52f8b-dd1d-446b-a090-c1e4157176e1","Type":"ContainerDied","Data":"2f649f29095ded77d082ff7b4aa4de9ed77717ad5b9e9734577faf5c03227b8f"} Feb 03 09:31:31 crc kubenswrapper[4756]: I0203 09:31:31.950607 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f649f29095ded77d082ff7b4aa4de9ed77717ad5b9e9734577faf5c03227b8f" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.056216 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.120367 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.121797 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.122793 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.135419 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-log-httpd\") pod \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.135771 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-config-data\") pod \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.135860 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-run-httpd\") pod \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.135935 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-sg-core-conf-yaml\") pod \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.135992 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3ca52f8b-dd1d-446b-a090-c1e4157176e1" (UID: "3ca52f8b-dd1d-446b-a090-c1e4157176e1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.136180 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3ca52f8b-dd1d-446b-a090-c1e4157176e1" (UID: "3ca52f8b-dd1d-446b-a090-c1e4157176e1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.136282 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-scripts\") pod \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.136372 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqhpp\" (UniqueName: \"kubernetes.io/projected/3ca52f8b-dd1d-446b-a090-c1e4157176e1-kube-api-access-dqhpp\") pod \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.136506 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-combined-ca-bundle\") pod \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\" (UID: \"3ca52f8b-dd1d-446b-a090-c1e4157176e1\") " Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.136964 4756 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.137044 4756 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ca52f8b-dd1d-446b-a090-c1e4157176e1-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.143674 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.161791 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca52f8b-dd1d-446b-a090-c1e4157176e1-kube-api-access-dqhpp" (OuterVolumeSpecName: "kube-api-access-dqhpp") pod "3ca52f8b-dd1d-446b-a090-c1e4157176e1" (UID: "3ca52f8b-dd1d-446b-a090-c1e4157176e1"). InnerVolumeSpecName "kube-api-access-dqhpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.166412 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-scripts" (OuterVolumeSpecName: "scripts") pod "3ca52f8b-dd1d-446b-a090-c1e4157176e1" (UID: "3ca52f8b-dd1d-446b-a090-c1e4157176e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.197008 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3ca52f8b-dd1d-446b-a090-c1e4157176e1" (UID: "3ca52f8b-dd1d-446b-a090-c1e4157176e1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.238928 4756 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.238962 4756 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-scripts\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.239094 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqhpp\" (UniqueName: \"kubernetes.io/projected/3ca52f8b-dd1d-446b-a090-c1e4157176e1-kube-api-access-dqhpp\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.250170 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ca52f8b-dd1d-446b-a090-c1e4157176e1" (UID: "3ca52f8b-dd1d-446b-a090-c1e4157176e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.291954 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-config-data" (OuterVolumeSpecName: "config-data") pod "3ca52f8b-dd1d-446b-a090-c1e4157176e1" (UID: "3ca52f8b-dd1d-446b-a090-c1e4157176e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.340640 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.340678 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca52f8b-dd1d-446b-a090-c1e4157176e1-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.972386 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.972554 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 03 09:31:32 crc kubenswrapper[4756]: I0203 09:31:32.979314 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.034509 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.040053 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.084951 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:31:33 crc kubenswrapper[4756]: E0203 09:31:33.085418 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="proxy-httpd" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.085459 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="proxy-httpd" Feb 03 09:31:33 crc kubenswrapper[4756]: E0203 09:31:33.085486 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="sg-core" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.085495 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="sg-core" Feb 03 09:31:33 crc kubenswrapper[4756]: E0203 09:31:33.085624 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="ceilometer-central-agent" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.085638 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="ceilometer-central-agent" Feb 03 09:31:33 crc kubenswrapper[4756]: E0203 09:31:33.085659 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="ceilometer-notification-agent" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.085666 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="ceilometer-notification-agent" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.085844 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="ceilometer-notification-agent" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.085866 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="sg-core" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.085872 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="ceilometer-central-agent" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.085892 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" containerName="proxy-httpd" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.087760 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.093005 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.093292 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.093922 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.118189 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.152824 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/198484fd-46b3-4299-8938-f49fdfdcacb0-run-httpd\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.152964 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhf9s\" (UniqueName: \"kubernetes.io/projected/198484fd-46b3-4299-8938-f49fdfdcacb0-kube-api-access-rhf9s\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.153091 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.153136 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.153176 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/198484fd-46b3-4299-8938-f49fdfdcacb0-log-httpd\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.153207 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-scripts\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.153226 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.153253 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-config-data\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255283 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255346 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/198484fd-46b3-4299-8938-f49fdfdcacb0-log-httpd\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255375 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255390 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-scripts\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255408 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-config-data\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255473 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/198484fd-46b3-4299-8938-f49fdfdcacb0-run-httpd\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255519 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhf9s\" (UniqueName: \"kubernetes.io/projected/198484fd-46b3-4299-8938-f49fdfdcacb0-kube-api-access-rhf9s\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255602 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.255950 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/198484fd-46b3-4299-8938-f49fdfdcacb0-log-httpd\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.256350 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/198484fd-46b3-4299-8938-f49fdfdcacb0-run-httpd\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.260611 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.260836 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-config-data\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.263956 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-scripts\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.265066 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.281966 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/198484fd-46b3-4299-8938-f49fdfdcacb0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.283811 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhf9s\" (UniqueName: \"kubernetes.io/projected/198484fd-46b3-4299-8938-f49fdfdcacb0-kube-api-access-rhf9s\") pod \"ceilometer-0\" (UID: \"198484fd-46b3-4299-8938-f49fdfdcacb0\") " pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.416315 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.624923 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca52f8b-dd1d-446b-a090-c1e4157176e1" path="/var/lib/kubelet/pods/3ca52f8b-dd1d-446b-a090-c1e4157176e1/volumes" Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.877224 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 03 09:31:33 crc kubenswrapper[4756]: W0203 09:31:33.879115 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod198484fd_46b3_4299_8938_f49fdfdcacb0.slice/crio-b3f75533ca268d71ed03192dc2914709d35d5feae18cf84cfb0b301914a32c19 WatchSource:0}: Error finding container b3f75533ca268d71ed03192dc2914709d35d5feae18cf84cfb0b301914a32c19: Status 404 returned error can't find the container with id b3f75533ca268d71ed03192dc2914709d35d5feae18cf84cfb0b301914a32c19 Feb 03 09:31:33 crc kubenswrapper[4756]: I0203 09:31:33.981622 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"198484fd-46b3-4299-8938-f49fdfdcacb0","Type":"ContainerStarted","Data":"b3f75533ca268d71ed03192dc2914709d35d5feae18cf84cfb0b301914a32c19"} Feb 03 09:31:34 crc kubenswrapper[4756]: I0203 09:31:34.993865 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"198484fd-46b3-4299-8938-f49fdfdcacb0","Type":"ContainerStarted","Data":"ce406beec7b73d64150bc6a8b51b2da4d07479ea3d2a315059bd568ad28131f6"} Feb 03 09:31:36 crc kubenswrapper[4756]: I0203 09:31:36.006884 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"198484fd-46b3-4299-8938-f49fdfdcacb0","Type":"ContainerStarted","Data":"a999228a6ef571da283fec26c0e610f188637ce3f48905c31bcb9e6a07c88060"} Feb 03 09:31:36 crc kubenswrapper[4756]: I0203 09:31:36.007305 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"198484fd-46b3-4299-8938-f49fdfdcacb0","Type":"ContainerStarted","Data":"94248dc4f403a46aa89a88fa8f5fca0f68e7b93cafad961c3882177f8795a638"} Feb 03 09:31:38 crc kubenswrapper[4756]: I0203 09:31:38.334606 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 03 09:31:39 crc kubenswrapper[4756]: I0203 09:31:39.033169 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"198484fd-46b3-4299-8938-f49fdfdcacb0","Type":"ContainerStarted","Data":"ac463347465e3aaa98d8e5b6947d6ec367b7b5d55c3d6bd78dfcf75a7b97446f"} Feb 03 09:31:39 crc kubenswrapper[4756]: I0203 09:31:39.062657 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.074520217 podStartE2EDuration="6.062637048s" podCreationTimestamp="2026-02-03 09:31:33 +0000 UTC" firstStartedPulling="2026-02-03 09:31:33.881103788 +0000 UTC m=+1285.031571153" lastFinishedPulling="2026-02-03 09:31:37.869220609 +0000 UTC m=+1289.019687984" observedRunningTime="2026-02-03 09:31:39.052994477 +0000 UTC m=+1290.203461872" watchObservedRunningTime="2026-02-03 09:31:39.062637048 +0000 UTC m=+1290.213104423" Feb 03 09:31:40 crc kubenswrapper[4756]: I0203 09:31:40.039902 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 03 09:31:43 crc kubenswrapper[4756]: I0203 09:31:43.566718 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:31:43 crc kubenswrapper[4756]: I0203 09:31:43.567113 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:32:03 crc kubenswrapper[4756]: I0203 09:32:03.426127 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 03 09:32:12 crc kubenswrapper[4756]: I0203 09:32:12.507094 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:32:13 crc kubenswrapper[4756]: I0203 09:32:13.371670 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:32:13 crc kubenswrapper[4756]: I0203 09:32:13.565731 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:32:13 crc kubenswrapper[4756]: I0203 09:32:13.565789 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:32:13 crc kubenswrapper[4756]: I0203 09:32:13.565836 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:32:13 crc kubenswrapper[4756]: I0203 09:32:13.566726 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"98d1f3cc0f9ab3d41cdcdd125bb9227106916f474a34592b5e034d6f0c2ba1ac"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:32:13 crc kubenswrapper[4756]: I0203 09:32:13.566780 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://98d1f3cc0f9ab3d41cdcdd125bb9227106916f474a34592b5e034d6f0c2ba1ac" gracePeriod=600 Feb 03 09:32:14 crc kubenswrapper[4756]: I0203 09:32:14.457940 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="98d1f3cc0f9ab3d41cdcdd125bb9227106916f474a34592b5e034d6f0c2ba1ac" exitCode=0 Feb 03 09:32:14 crc kubenswrapper[4756]: I0203 09:32:14.457976 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"98d1f3cc0f9ab3d41cdcdd125bb9227106916f474a34592b5e034d6f0c2ba1ac"} Feb 03 09:32:14 crc kubenswrapper[4756]: I0203 09:32:14.458426 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437"} Feb 03 09:32:14 crc kubenswrapper[4756]: I0203 09:32:14.458460 4756 scope.go:117] "RemoveContainer" containerID="0da902e4b7c741d00f45d02b231b9b641521e6c5b54725cc76fd568b0058caf7" Feb 03 09:32:16 crc kubenswrapper[4756]: I0203 09:32:16.809410 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerName="rabbitmq" containerID="cri-o://fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4" gracePeriod=604796 Feb 03 09:32:17 crc kubenswrapper[4756]: I0203 09:32:17.443723 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerName="rabbitmq" containerID="cri-o://1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f" gracePeriod=604796 Feb 03 09:32:20 crc kubenswrapper[4756]: I0203 09:32:20.687931 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Feb 03 09:32:21 crc kubenswrapper[4756]: I0203 09:32:21.019106 4756 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.417480 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.501651 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-plugins\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.501761 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.501909 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-plugins-conf\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.501985 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-config-data\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.502083 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlwzr\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-kube-api-access-xlwzr\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.502347 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.503317 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.503475 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-pod-info\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.503584 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-erlang-cookie-secret\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.505728 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-confd\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.505916 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-server-conf\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.506010 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-tls\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.506174 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-erlang-cookie\") pod \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\" (UID: \"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c\") " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.507743 4756 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.507958 4756 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.508678 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-kube-api-access-xlwzr" (OuterVolumeSpecName: "kube-api-access-xlwzr") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "kube-api-access-xlwzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.509079 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.512553 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.515029 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-pod-info" (OuterVolumeSpecName: "pod-info") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.516125 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.518181 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.537143 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-config-data" (OuterVolumeSpecName: "config-data") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.583970 4756 generic.go:334] "Generic (PLEG): container finished" podID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerID="fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4" exitCode=0 Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.584028 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c","Type":"ContainerDied","Data":"fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4"} Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.584060 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9db7c10a-3acf-423c-8dcf-3bcd7b832b5c","Type":"ContainerDied","Data":"8feab9b6921feea3dc1c80e8e6eeb237b6abeb956563b87b52840f8dda6a8509"} Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.584083 4756 scope.go:117] "RemoveContainer" containerID="fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.584132 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.600273 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-server-conf" (OuterVolumeSpecName: "server-conf") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.610374 4756 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.610411 4756 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.610441 4756 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.610453 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.610476 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlwzr\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-kube-api-access-xlwzr\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.610488 4756 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-pod-info\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.610498 4756 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.610508 4756 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-server-conf\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.658665 4756 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.675528 4756 scope.go:117] "RemoveContainer" containerID="ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.683865 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" (UID: "9db7c10a-3acf-423c-8dcf-3bcd7b832b5c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.706962 4756 scope.go:117] "RemoveContainer" containerID="fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4" Feb 03 09:32:23 crc kubenswrapper[4756]: E0203 09:32:23.707522 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4\": container with ID starting with fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4 not found: ID does not exist" containerID="fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.707581 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4"} err="failed to get container status \"fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4\": rpc error: code = NotFound desc = could not find container \"fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4\": container with ID starting with fa9048eec30604142e71b6021e37a14fff4ecab9f7086f990f283839c657fdd4 not found: ID does not exist" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.707626 4756 scope.go:117] "RemoveContainer" containerID="ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f" Feb 03 09:32:23 crc kubenswrapper[4756]: E0203 09:32:23.707959 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f\": container with ID starting with ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f not found: ID does not exist" containerID="ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.707986 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f"} err="failed to get container status \"ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f\": rpc error: code = NotFound desc = could not find container \"ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f\": container with ID starting with ffbea8afce31cd687ede0ecbce59c5e19c223b2aeb1feba5c8cbae6309c0900f not found: ID does not exist" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.713222 4756 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.713242 4756 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.932910 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.936056 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.947065 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.958655 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:32:23 crc kubenswrapper[4756]: E0203 09:32:23.959047 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerName="setup-container" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.959064 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerName="setup-container" Feb 03 09:32:23 crc kubenswrapper[4756]: E0203 09:32:23.959082 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerName="setup-container" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.959088 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerName="setup-container" Feb 03 09:32:23 crc kubenswrapper[4756]: E0203 09:32:23.959101 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerName="rabbitmq" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.959109 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerName="rabbitmq" Feb 03 09:32:23 crc kubenswrapper[4756]: E0203 09:32:23.959131 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerName="rabbitmq" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.959138 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerName="rabbitmq" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.959298 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" containerName="rabbitmq" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.959314 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerName="rabbitmq" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.960236 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.963023 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.966603 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.966894 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.967119 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.967361 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.967409 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.967545 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hr92v" Feb 03 09:32:23 crc kubenswrapper[4756]: I0203 09:32:23.995307 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.020816 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-server-conf\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.020918 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-plugins-conf\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.020955 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-confd\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021012 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021097 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-tls\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021145 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-erlang-cookie\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021208 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-pod-info\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021238 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-plugins\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021278 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-config-data\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021300 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-erlang-cookie-secret\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021328 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72vpb\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-kube-api-access-72vpb\") pod \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\" (UID: \"f2aeafe0-b129-44d5-86d9-0a95a7ac180d\") " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021610 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmh9k\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-kube-api-access-vmh9k\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021649 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021688 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021713 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-config-data\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021749 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021872 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021912 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021943 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f76f412-ff7c-4857-9670-4721dc5c3080-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.021976 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f76f412-ff7c-4857-9670-4721dc5c3080-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.022007 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.022053 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.032317 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.032813 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.033789 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.037349 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-pod-info" (OuterVolumeSpecName: "pod-info") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.037830 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.037876 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.041702 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-kube-api-access-72vpb" (OuterVolumeSpecName: "kube-api-access-72vpb") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "kube-api-access-72vpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.051854 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.071571 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-config-data" (OuterVolumeSpecName: "config-data") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.087331 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-server-conf" (OuterVolumeSpecName: "server-conf") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.124705 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.124969 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125045 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f76f412-ff7c-4857-9670-4721dc5c3080-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125185 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f76f412-ff7c-4857-9670-4721dc5c3080-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125262 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125319 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125385 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmh9k\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-kube-api-access-vmh9k\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125419 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125466 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125488 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-config-data\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125531 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125662 4756 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125684 4756 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125697 4756 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-pod-info\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125708 4756 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125719 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125731 4756 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125743 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72vpb\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-kube-api-access-72vpb\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125754 4756 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-server-conf\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125766 4756 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.125788 4756 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.127002 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.127268 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.127933 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-config-data\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.128180 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.128273 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.128765 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.129981 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f76f412-ff7c-4857-9670-4721dc5c3080-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.131363 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.131641 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f76f412-ff7c-4857-9670-4721dc5c3080-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.136814 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f76f412-ff7c-4857-9670-4721dc5c3080-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.148199 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmh9k\" (UniqueName: \"kubernetes.io/projected/5f76f412-ff7c-4857-9670-4721dc5c3080-kube-api-access-vmh9k\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.165287 4756 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.172386 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"5f76f412-ff7c-4857-9670-4721dc5c3080\") " pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.194065 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f2aeafe0-b129-44d5-86d9-0a95a7ac180d" (UID: "f2aeafe0-b129-44d5-86d9-0a95a7ac180d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.227177 4756 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2aeafe0-b129-44d5-86d9-0a95a7ac180d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.227220 4756 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.289961 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.600449 4756 generic.go:334] "Generic (PLEG): container finished" podID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" containerID="1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f" exitCode=0 Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.600484 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f2aeafe0-b129-44d5-86d9-0a95a7ac180d","Type":"ContainerDied","Data":"1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f"} Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.600556 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f2aeafe0-b129-44d5-86d9-0a95a7ac180d","Type":"ContainerDied","Data":"1eca17f1e9bb4fb18827ebe59224016c1b128d8aa09d0a8610ed79914da98851"} Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.600570 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.600584 4756 scope.go:117] "RemoveContainer" containerID="1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.621225 4756 scope.go:117] "RemoveContainer" containerID="30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.653174 4756 scope.go:117] "RemoveContainer" containerID="1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.653272 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:32:24 crc kubenswrapper[4756]: E0203 09:32:24.653878 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f\": container with ID starting with 1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f not found: ID does not exist" containerID="1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.653913 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f"} err="failed to get container status \"1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f\": rpc error: code = NotFound desc = could not find container \"1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f\": container with ID starting with 1ace44f5d6ea1e85743403932f0b879ef33d3fd5983fdd412346626cebd8583f not found: ID does not exist" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.653933 4756 scope.go:117] "RemoveContainer" containerID="30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53" Feb 03 09:32:24 crc kubenswrapper[4756]: E0203 09:32:24.654688 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53\": container with ID starting with 30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53 not found: ID does not exist" containerID="30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.654742 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53"} err="failed to get container status \"30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53\": rpc error: code = NotFound desc = could not find container \"30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53\": container with ID starting with 30c25321645ea75f5144cbbcc3a66b038ebfba49dcc7874b5da5f035bebedf53 not found: ID does not exist" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.661975 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.671244 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.673159 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.680219 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.680555 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.680724 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.680924 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.681091 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.681126 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qzgw2" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.681336 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.686879 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.740745 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.741091 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60b10ab9-d28a-4c83-9e67-02d1499685ce-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.741252 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.741364 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.741479 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60b10ab9-d28a-4c83-9e67-02d1499685ce-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.741604 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.743764 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.743882 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.744107 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.744233 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvh8f\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-kube-api-access-tvh8f\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.744380 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.777525 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846336 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846411 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60b10ab9-d28a-4c83-9e67-02d1499685ce-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846495 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846542 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846568 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60b10ab9-d28a-4c83-9e67-02d1499685ce-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846591 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846640 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846660 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846757 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846815 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvh8f\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-kube-api-access-tvh8f\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.846841 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.847368 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.847569 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.847574 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.847951 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.848276 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.848790 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/60b10ab9-d28a-4c83-9e67-02d1499685ce-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.850172 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/60b10ab9-d28a-4c83-9e67-02d1499685ce-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.852577 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.856414 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/60b10ab9-d28a-4c83-9e67-02d1499685ce-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.856529 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.867538 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvh8f\" (UniqueName: \"kubernetes.io/projected/60b10ab9-d28a-4c83-9e67-02d1499685ce-kube-api-access-tvh8f\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:24 crc kubenswrapper[4756]: I0203 09:32:24.882224 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"60b10ab9-d28a-4c83-9e67-02d1499685ce\") " pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.011263 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:32:25 crc kubenswrapper[4756]: W0203 09:32:25.498785 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60b10ab9_d28a_4c83_9e67_02d1499685ce.slice/crio-33b3a6486dc4832e7bde4cb699926fb78059d500a383ff499631b7c13a532f3f WatchSource:0}: Error finding container 33b3a6486dc4832e7bde4cb699926fb78059d500a383ff499631b7c13a532f3f: Status 404 returned error can't find the container with id 33b3a6486dc4832e7bde4cb699926fb78059d500a383ff499631b7c13a532f3f Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.499076 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.542514 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-vx9g6"] Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.544768 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.547042 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.566021 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-vx9g6"] Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.626741 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9db7c10a-3acf-423c-8dcf-3bcd7b832b5c" path="/var/lib/kubelet/pods/9db7c10a-3acf-423c-8dcf-3bcd7b832b5c/volumes" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.627719 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2aeafe0-b129-44d5-86d9-0a95a7ac180d" path="/var/lib/kubelet/pods/f2aeafe0-b129-44d5-86d9-0a95a7ac180d/volumes" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.628258 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"60b10ab9-d28a-4c83-9e67-02d1499685ce","Type":"ContainerStarted","Data":"33b3a6486dc4832e7bde4cb699926fb78059d500a383ff499631b7c13a532f3f"} Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.628283 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f76f412-ff7c-4857-9670-4721dc5c3080","Type":"ContainerStarted","Data":"49f1750ecdf49bc1c5cf2390b68a7b1800545db9ba7b99d62c078b2a863d407d"} Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.662934 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.663017 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72jrw\" (UniqueName: \"kubernetes.io/projected/7acf11c3-6d46-400c-8089-39129049e082-kube-api-access-72jrw\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.663048 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-config\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.663069 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.663413 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.663502 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-svc\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.663533 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.765184 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.765257 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72jrw\" (UniqueName: \"kubernetes.io/projected/7acf11c3-6d46-400c-8089-39129049e082-kube-api-access-72jrw\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.765287 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-config\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.765308 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.765398 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.765425 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-svc\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.765444 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.766331 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.766347 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.766418 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-config\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.766460 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.766568 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-svc\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.767086 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:25 crc kubenswrapper[4756]: I0203 09:32:25.785125 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72jrw\" (UniqueName: \"kubernetes.io/projected/7acf11c3-6d46-400c-8089-39129049e082-kube-api-access-72jrw\") pod \"dnsmasq-dns-d558885bc-vx9g6\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:26 crc kubenswrapper[4756]: I0203 09:32:26.063108 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:26 crc kubenswrapper[4756]: I0203 09:32:26.628850 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-vx9g6"] Feb 03 09:32:26 crc kubenswrapper[4756]: W0203 09:32:26.630139 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7acf11c3_6d46_400c_8089_39129049e082.slice/crio-0218c437216e6a56fd854a28c427c996c35ce1efe9d38c2748dd39721a6ac4ca WatchSource:0}: Error finding container 0218c437216e6a56fd854a28c427c996c35ce1efe9d38c2748dd39721a6ac4ca: Status 404 returned error can't find the container with id 0218c437216e6a56fd854a28c427c996c35ce1efe9d38c2748dd39721a6ac4ca Feb 03 09:32:26 crc kubenswrapper[4756]: I0203 09:32:26.630374 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f76f412-ff7c-4857-9670-4721dc5c3080","Type":"ContainerStarted","Data":"4e1ae88a96f3ed91897e5d4ee6d2022f0181342e8008c05e99ba707cc5d72042"} Feb 03 09:32:27 crc kubenswrapper[4756]: I0203 09:32:27.639279 4756 generic.go:334] "Generic (PLEG): container finished" podID="7acf11c3-6d46-400c-8089-39129049e082" containerID="5520cb10ba7a81f41de98ba5ce9e740fda20749432f80f3a3a8d5319bc8dbf19" exitCode=0 Feb 03 09:32:27 crc kubenswrapper[4756]: I0203 09:32:27.639328 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" event={"ID":"7acf11c3-6d46-400c-8089-39129049e082","Type":"ContainerDied","Data":"5520cb10ba7a81f41de98ba5ce9e740fda20749432f80f3a3a8d5319bc8dbf19"} Feb 03 09:32:27 crc kubenswrapper[4756]: I0203 09:32:27.639615 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" event={"ID":"7acf11c3-6d46-400c-8089-39129049e082","Type":"ContainerStarted","Data":"0218c437216e6a56fd854a28c427c996c35ce1efe9d38c2748dd39721a6ac4ca"} Feb 03 09:32:27 crc kubenswrapper[4756]: I0203 09:32:27.648055 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"60b10ab9-d28a-4c83-9e67-02d1499685ce","Type":"ContainerStarted","Data":"9969684674eaecbe49cf888a9d28011aa5f4a17cb875ec5baea30af437bc79f5"} Feb 03 09:32:28 crc kubenswrapper[4756]: I0203 09:32:28.658020 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" event={"ID":"7acf11c3-6d46-400c-8089-39129049e082","Type":"ContainerStarted","Data":"5088c9df342ca9460e529c26cc3e3162cf3ac5a05ed7e33bc5b31d07f22abdca"} Feb 03 09:32:28 crc kubenswrapper[4756]: I0203 09:32:28.658980 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:28 crc kubenswrapper[4756]: I0203 09:32:28.682041 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" podStartSLOduration=3.682015548 podStartE2EDuration="3.682015548s" podCreationTimestamp="2026-02-03 09:32:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:32:28.676838687 +0000 UTC m=+1339.827306062" watchObservedRunningTime="2026-02-03 09:32:28.682015548 +0000 UTC m=+1339.832482963" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.065169 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.121338 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-qlqfx"] Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.121640 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" podUID="97678a5a-ea1a-477f-8a7f-adff844186a4" containerName="dnsmasq-dns" containerID="cri-o://cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5" gracePeriod=10 Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.306737 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-wtq86"] Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.324400 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-wtq86"] Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.324784 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.353153 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.353215 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-config\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.353279 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.353307 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw4p5\" (UniqueName: \"kubernetes.io/projected/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-kube-api-access-cw4p5\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.353330 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.353371 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.353396 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.454148 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.454206 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.455077 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.455376 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.455512 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.456143 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.456286 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-config\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.456411 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.456478 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw4p5\" (UniqueName: \"kubernetes.io/projected/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-kube-api-access-cw4p5\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.456509 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.457170 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.457750 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-config\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.458266 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.478315 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw4p5\" (UniqueName: \"kubernetes.io/projected/2fa5e86f-f1ee-4e6e-9dba-26c7055a767e-kube-api-access-cw4p5\") pod \"dnsmasq-dns-78c64bc9c5-wtq86\" (UID: \"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e\") " pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.645701 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.659165 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-svc\") pod \"97678a5a-ea1a-477f-8a7f-adff844186a4\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.659330 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-swift-storage-0\") pod \"97678a5a-ea1a-477f-8a7f-adff844186a4\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.659412 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-nb\") pod \"97678a5a-ea1a-477f-8a7f-adff844186a4\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.659463 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6kzb\" (UniqueName: \"kubernetes.io/projected/97678a5a-ea1a-477f-8a7f-adff844186a4-kube-api-access-p6kzb\") pod \"97678a5a-ea1a-477f-8a7f-adff844186a4\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.659486 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-config\") pod \"97678a5a-ea1a-477f-8a7f-adff844186a4\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.659548 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-sb\") pod \"97678a5a-ea1a-477f-8a7f-adff844186a4\" (UID: \"97678a5a-ea1a-477f-8a7f-adff844186a4\") " Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.660703 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.689761 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97678a5a-ea1a-477f-8a7f-adff844186a4-kube-api-access-p6kzb" (OuterVolumeSpecName: "kube-api-access-p6kzb") pod "97678a5a-ea1a-477f-8a7f-adff844186a4" (UID: "97678a5a-ea1a-477f-8a7f-adff844186a4"). InnerVolumeSpecName "kube-api-access-p6kzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.755234 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "97678a5a-ea1a-477f-8a7f-adff844186a4" (UID: "97678a5a-ea1a-477f-8a7f-adff844186a4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.764514 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "97678a5a-ea1a-477f-8a7f-adff844186a4" (UID: "97678a5a-ea1a-477f-8a7f-adff844186a4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.770614 4756 generic.go:334] "Generic (PLEG): container finished" podID="97678a5a-ea1a-477f-8a7f-adff844186a4" containerID="cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5" exitCode=0 Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.770670 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" event={"ID":"97678a5a-ea1a-477f-8a7f-adff844186a4","Type":"ContainerDied","Data":"cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5"} Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.770696 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" event={"ID":"97678a5a-ea1a-477f-8a7f-adff844186a4","Type":"ContainerDied","Data":"ff9f147fa6cfe2691603a6204cb5028acb330b608fd84f21785568e3a8134e25"} Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.770712 4756 scope.go:117] "RemoveContainer" containerID="cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.770870 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-qlqfx" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.778967 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.779008 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6kzb\" (UniqueName: \"kubernetes.io/projected/97678a5a-ea1a-477f-8a7f-adff844186a4-kube-api-access-p6kzb\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.779025 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.781021 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "97678a5a-ea1a-477f-8a7f-adff844186a4" (UID: "97678a5a-ea1a-477f-8a7f-adff844186a4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.822043 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "97678a5a-ea1a-477f-8a7f-adff844186a4" (UID: "97678a5a-ea1a-477f-8a7f-adff844186a4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.834393 4756 scope.go:117] "RemoveContainer" containerID="ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.834752 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-config" (OuterVolumeSpecName: "config") pod "97678a5a-ea1a-477f-8a7f-adff844186a4" (UID: "97678a5a-ea1a-477f-8a7f-adff844186a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.869724 4756 scope.go:117] "RemoveContainer" containerID="cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5" Feb 03 09:32:36 crc kubenswrapper[4756]: E0203 09:32:36.870113 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5\": container with ID starting with cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5 not found: ID does not exist" containerID="cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.870151 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5"} err="failed to get container status \"cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5\": rpc error: code = NotFound desc = could not find container \"cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5\": container with ID starting with cbf6d6d37e857316f1c797767abd3f797083509d4344acbc0f42fe1dbcfa8dc5 not found: ID does not exist" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.870174 4756 scope.go:117] "RemoveContainer" containerID="ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f" Feb 03 09:32:36 crc kubenswrapper[4756]: E0203 09:32:36.870370 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f\": container with ID starting with ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f not found: ID does not exist" containerID="ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.870401 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f"} err="failed to get container status \"ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f\": rpc error: code = NotFound desc = could not find container \"ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f\": container with ID starting with ad694a65fec90c222f10e86cfad2e8337301abe13557b31c979d5ba7ec57e17f not found: ID does not exist" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.882152 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.882186 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:36 crc kubenswrapper[4756]: I0203 09:32:36.882199 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97678a5a-ea1a-477f-8a7f-adff844186a4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:37 crc kubenswrapper[4756]: I0203 09:32:37.103590 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-qlqfx"] Feb 03 09:32:37 crc kubenswrapper[4756]: I0203 09:32:37.115913 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-qlqfx"] Feb 03 09:32:37 crc kubenswrapper[4756]: I0203 09:32:37.219730 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-wtq86"] Feb 03 09:32:37 crc kubenswrapper[4756]: W0203 09:32:37.220800 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fa5e86f_f1ee_4e6e_9dba_26c7055a767e.slice/crio-8527d0ba8168e9d9c265f92bbfafca93a8db05189216e6530cb615c63ad20d8f WatchSource:0}: Error finding container 8527d0ba8168e9d9c265f92bbfafca93a8db05189216e6530cb615c63ad20d8f: Status 404 returned error can't find the container with id 8527d0ba8168e9d9c265f92bbfafca93a8db05189216e6530cb615c63ad20d8f Feb 03 09:32:37 crc kubenswrapper[4756]: I0203 09:32:37.627246 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97678a5a-ea1a-477f-8a7f-adff844186a4" path="/var/lib/kubelet/pods/97678a5a-ea1a-477f-8a7f-adff844186a4/volumes" Feb 03 09:32:37 crc kubenswrapper[4756]: I0203 09:32:37.782858 4756 generic.go:334] "Generic (PLEG): container finished" podID="2fa5e86f-f1ee-4e6e-9dba-26c7055a767e" containerID="2f14cad2f1f7006ba9664dce6312f4de25b8ca261107a7eacc2f846d4096de91" exitCode=0 Feb 03 09:32:37 crc kubenswrapper[4756]: I0203 09:32:37.782965 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" event={"ID":"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e","Type":"ContainerDied","Data":"2f14cad2f1f7006ba9664dce6312f4de25b8ca261107a7eacc2f846d4096de91"} Feb 03 09:32:37 crc kubenswrapper[4756]: I0203 09:32:37.783301 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" event={"ID":"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e","Type":"ContainerStarted","Data":"8527d0ba8168e9d9c265f92bbfafca93a8db05189216e6530cb615c63ad20d8f"} Feb 03 09:32:38 crc kubenswrapper[4756]: I0203 09:32:38.795184 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" event={"ID":"2fa5e86f-f1ee-4e6e-9dba-26c7055a767e","Type":"ContainerStarted","Data":"3f192ed0438272bff419354904ea2aead3aef2e3a603a3922b63834dfa070c6d"} Feb 03 09:32:38 crc kubenswrapper[4756]: I0203 09:32:38.795435 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:38 crc kubenswrapper[4756]: I0203 09:32:38.823655 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" podStartSLOduration=2.823637647 podStartE2EDuration="2.823637647s" podCreationTimestamp="2026-02-03 09:32:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:32:38.8154129 +0000 UTC m=+1349.965880285" watchObservedRunningTime="2026-02-03 09:32:38.823637647 +0000 UTC m=+1349.974105022" Feb 03 09:32:46 crc kubenswrapper[4756]: I0203 09:32:46.662665 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-wtq86" Feb 03 09:32:46 crc kubenswrapper[4756]: I0203 09:32:46.724278 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-vx9g6"] Feb 03 09:32:46 crc kubenswrapper[4756]: I0203 09:32:46.724628 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" podUID="7acf11c3-6d46-400c-8089-39129049e082" containerName="dnsmasq-dns" containerID="cri-o://5088c9df342ca9460e529c26cc3e3162cf3ac5a05ed7e33bc5b31d07f22abdca" gracePeriod=10 Feb 03 09:32:46 crc kubenswrapper[4756]: I0203 09:32:46.866866 4756 generic.go:334] "Generic (PLEG): container finished" podID="7acf11c3-6d46-400c-8089-39129049e082" containerID="5088c9df342ca9460e529c26cc3e3162cf3ac5a05ed7e33bc5b31d07f22abdca" exitCode=0 Feb 03 09:32:46 crc kubenswrapper[4756]: I0203 09:32:46.866927 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" event={"ID":"7acf11c3-6d46-400c-8089-39129049e082","Type":"ContainerDied","Data":"5088c9df342ca9460e529c26cc3e3162cf3ac5a05ed7e33bc5b31d07f22abdca"} Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.200091 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.282271 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72jrw\" (UniqueName: \"kubernetes.io/projected/7acf11c3-6d46-400c-8089-39129049e082-kube-api-access-72jrw\") pod \"7acf11c3-6d46-400c-8089-39129049e082\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.282360 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-config\") pod \"7acf11c3-6d46-400c-8089-39129049e082\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.282555 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-svc\") pod \"7acf11c3-6d46-400c-8089-39129049e082\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.283476 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-sb\") pod \"7acf11c3-6d46-400c-8089-39129049e082\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.283576 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-nb\") pod \"7acf11c3-6d46-400c-8089-39129049e082\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.283627 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-swift-storage-0\") pod \"7acf11c3-6d46-400c-8089-39129049e082\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.283676 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-openstack-edpm-ipam\") pod \"7acf11c3-6d46-400c-8089-39129049e082\" (UID: \"7acf11c3-6d46-400c-8089-39129049e082\") " Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.288000 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7acf11c3-6d46-400c-8089-39129049e082-kube-api-access-72jrw" (OuterVolumeSpecName: "kube-api-access-72jrw") pod "7acf11c3-6d46-400c-8089-39129049e082" (UID: "7acf11c3-6d46-400c-8089-39129049e082"). InnerVolumeSpecName "kube-api-access-72jrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.331587 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7acf11c3-6d46-400c-8089-39129049e082" (UID: "7acf11c3-6d46-400c-8089-39129049e082"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.337151 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7acf11c3-6d46-400c-8089-39129049e082" (UID: "7acf11c3-6d46-400c-8089-39129049e082"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.342970 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7acf11c3-6d46-400c-8089-39129049e082" (UID: "7acf11c3-6d46-400c-8089-39129049e082"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.345687 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7acf11c3-6d46-400c-8089-39129049e082" (UID: "7acf11c3-6d46-400c-8089-39129049e082"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.350579 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "7acf11c3-6d46-400c-8089-39129049e082" (UID: "7acf11c3-6d46-400c-8089-39129049e082"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.352171 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-config" (OuterVolumeSpecName: "config") pod "7acf11c3-6d46-400c-8089-39129049e082" (UID: "7acf11c3-6d46-400c-8089-39129049e082"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.386718 4756 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.386768 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.386782 4756 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.386794 4756 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.386807 4756 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.386821 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72jrw\" (UniqueName: \"kubernetes.io/projected/7acf11c3-6d46-400c-8089-39129049e082-kube-api-access-72jrw\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.386834 4756 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7acf11c3-6d46-400c-8089-39129049e082-config\") on node \"crc\" DevicePath \"\"" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.876467 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" event={"ID":"7acf11c3-6d46-400c-8089-39129049e082","Type":"ContainerDied","Data":"0218c437216e6a56fd854a28c427c996c35ce1efe9d38c2748dd39721a6ac4ca"} Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.876526 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-vx9g6" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.876532 4756 scope.go:117] "RemoveContainer" containerID="5088c9df342ca9460e529c26cc3e3162cf3ac5a05ed7e33bc5b31d07f22abdca" Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.896033 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-vx9g6"] Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.904208 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-vx9g6"] Feb 03 09:32:47 crc kubenswrapper[4756]: I0203 09:32:47.914302 4756 scope.go:117] "RemoveContainer" containerID="5520cb10ba7a81f41de98ba5ce9e740fda20749432f80f3a3a8d5319bc8dbf19" Feb 03 09:32:49 crc kubenswrapper[4756]: I0203 09:32:49.626539 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7acf11c3-6d46-400c-8089-39129049e082" path="/var/lib/kubelet/pods/7acf11c3-6d46-400c-8089-39129049e082/volumes" Feb 03 09:32:55 crc kubenswrapper[4756]: I0203 09:32:55.996777 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v"] Feb 03 09:32:55 crc kubenswrapper[4756]: E0203 09:32:55.997762 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97678a5a-ea1a-477f-8a7f-adff844186a4" containerName="dnsmasq-dns" Feb 03 09:32:55 crc kubenswrapper[4756]: I0203 09:32:55.997777 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="97678a5a-ea1a-477f-8a7f-adff844186a4" containerName="dnsmasq-dns" Feb 03 09:32:55 crc kubenswrapper[4756]: E0203 09:32:55.997797 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97678a5a-ea1a-477f-8a7f-adff844186a4" containerName="init" Feb 03 09:32:55 crc kubenswrapper[4756]: I0203 09:32:55.997804 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="97678a5a-ea1a-477f-8a7f-adff844186a4" containerName="init" Feb 03 09:32:55 crc kubenswrapper[4756]: E0203 09:32:55.997824 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7acf11c3-6d46-400c-8089-39129049e082" containerName="init" Feb 03 09:32:55 crc kubenswrapper[4756]: I0203 09:32:55.997832 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7acf11c3-6d46-400c-8089-39129049e082" containerName="init" Feb 03 09:32:55 crc kubenswrapper[4756]: E0203 09:32:55.997861 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7acf11c3-6d46-400c-8089-39129049e082" containerName="dnsmasq-dns" Feb 03 09:32:55 crc kubenswrapper[4756]: I0203 09:32:55.997867 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7acf11c3-6d46-400c-8089-39129049e082" containerName="dnsmasq-dns" Feb 03 09:32:55 crc kubenswrapper[4756]: I0203 09:32:55.998024 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="97678a5a-ea1a-477f-8a7f-adff844186a4" containerName="dnsmasq-dns" Feb 03 09:32:55 crc kubenswrapper[4756]: I0203 09:32:55.998043 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="7acf11c3-6d46-400c-8089-39129049e082" containerName="dnsmasq-dns" Feb 03 09:32:55 crc kubenswrapper[4756]: I0203 09:32:55.998700 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.001153 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.001208 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.001387 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.014231 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v"] Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.015192 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.044693 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c7ph\" (UniqueName: \"kubernetes.io/projected/200f01ff-d33a-474b-944e-e6ed2486db68-kube-api-access-6c7ph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.044757 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.044811 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.044846 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.146774 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c7ph\" (UniqueName: \"kubernetes.io/projected/200f01ff-d33a-474b-944e-e6ed2486db68-kube-api-access-6c7ph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.146828 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.146869 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.146903 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.152804 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.154014 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.154643 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.184161 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c7ph\" (UniqueName: \"kubernetes.io/projected/200f01ff-d33a-474b-944e-e6ed2486db68-kube-api-access-6c7ph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.321609 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.852415 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v"] Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.862593 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 09:32:56 crc kubenswrapper[4756]: I0203 09:32:56.956032 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" event={"ID":"200f01ff-d33a-474b-944e-e6ed2486db68","Type":"ContainerStarted","Data":"44a7399946124b3c4a01fe9cdbf82a90d97b85587a0d8056bf60b6b30db3dd1d"} Feb 03 09:32:58 crc kubenswrapper[4756]: I0203 09:32:58.980630 4756 generic.go:334] "Generic (PLEG): container finished" podID="5f76f412-ff7c-4857-9670-4721dc5c3080" containerID="4e1ae88a96f3ed91897e5d4ee6d2022f0181342e8008c05e99ba707cc5d72042" exitCode=0 Feb 03 09:32:58 crc kubenswrapper[4756]: I0203 09:32:58.980883 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f76f412-ff7c-4857-9670-4721dc5c3080","Type":"ContainerDied","Data":"4e1ae88a96f3ed91897e5d4ee6d2022f0181342e8008c05e99ba707cc5d72042"} Feb 03 09:32:59 crc kubenswrapper[4756]: I0203 09:32:59.990185 4756 generic.go:334] "Generic (PLEG): container finished" podID="60b10ab9-d28a-4c83-9e67-02d1499685ce" containerID="9969684674eaecbe49cf888a9d28011aa5f4a17cb875ec5baea30af437bc79f5" exitCode=0 Feb 03 09:32:59 crc kubenswrapper[4756]: I0203 09:32:59.990258 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"60b10ab9-d28a-4c83-9e67-02d1499685ce","Type":"ContainerDied","Data":"9969684674eaecbe49cf888a9d28011aa5f4a17cb875ec5baea30af437bc79f5"} Feb 03 09:32:59 crc kubenswrapper[4756]: I0203 09:32:59.992227 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f76f412-ff7c-4857-9670-4721dc5c3080","Type":"ContainerStarted","Data":"4ea5f2af0942e9596af22bd60278f17a43296d8995e73b53140b725405130026"} Feb 03 09:32:59 crc kubenswrapper[4756]: I0203 09:32:59.992840 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 03 09:33:00 crc kubenswrapper[4756]: I0203 09:33:00.065978 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.065961278 podStartE2EDuration="37.065961278s" podCreationTimestamp="2026-02-03 09:32:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:33:00.056946476 +0000 UTC m=+1371.207413861" watchObservedRunningTime="2026-02-03 09:33:00.065961278 +0000 UTC m=+1371.216428653" Feb 03 09:33:01 crc kubenswrapper[4756]: I0203 09:33:01.011721 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"60b10ab9-d28a-4c83-9e67-02d1499685ce","Type":"ContainerStarted","Data":"400e3ad0f73a3f63bee853f7096c80d0ec836b132f668d06c93c36f2c32acce2"} Feb 03 09:33:01 crc kubenswrapper[4756]: I0203 09:33:01.012252 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:33:01 crc kubenswrapper[4756]: I0203 09:33:01.043098 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.043080027 podStartE2EDuration="37.043080027s" podCreationTimestamp="2026-02-03 09:32:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 09:33:01.033151395 +0000 UTC m=+1372.183618790" watchObservedRunningTime="2026-02-03 09:33:01.043080027 +0000 UTC m=+1372.193547402" Feb 03 09:33:10 crc kubenswrapper[4756]: I0203 09:33:10.109298 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" event={"ID":"200f01ff-d33a-474b-944e-e6ed2486db68","Type":"ContainerStarted","Data":"0c013c1b5c2434532471fe90c8af9bd418ed4fb894a6ab1d3febf26025ee5328"} Feb 03 09:33:10 crc kubenswrapper[4756]: I0203 09:33:10.131589 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" podStartSLOduration=2.525288233 podStartE2EDuration="15.131571857s" podCreationTimestamp="2026-02-03 09:32:55 +0000 UTC" firstStartedPulling="2026-02-03 09:32:56.86225736 +0000 UTC m=+1368.012724735" lastFinishedPulling="2026-02-03 09:33:09.468540974 +0000 UTC m=+1380.619008359" observedRunningTime="2026-02-03 09:33:10.123177953 +0000 UTC m=+1381.273645348" watchObservedRunningTime="2026-02-03 09:33:10.131571857 +0000 UTC m=+1381.282039232" Feb 03 09:33:14 crc kubenswrapper[4756]: I0203 09:33:14.292760 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 03 09:33:15 crc kubenswrapper[4756]: I0203 09:33:15.014715 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 03 09:33:21 crc kubenswrapper[4756]: I0203 09:33:21.213137 4756 generic.go:334] "Generic (PLEG): container finished" podID="200f01ff-d33a-474b-944e-e6ed2486db68" containerID="0c013c1b5c2434532471fe90c8af9bd418ed4fb894a6ab1d3febf26025ee5328" exitCode=0 Feb 03 09:33:21 crc kubenswrapper[4756]: I0203 09:33:21.213238 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" event={"ID":"200f01ff-d33a-474b-944e-e6ed2486db68","Type":"ContainerDied","Data":"0c013c1b5c2434532471fe90c8af9bd418ed4fb894a6ab1d3febf26025ee5328"} Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.658866 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.752050 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-inventory\") pod \"200f01ff-d33a-474b-944e-e6ed2486db68\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.752177 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c7ph\" (UniqueName: \"kubernetes.io/projected/200f01ff-d33a-474b-944e-e6ed2486db68-kube-api-access-6c7ph\") pod \"200f01ff-d33a-474b-944e-e6ed2486db68\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.752237 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-ssh-key-openstack-edpm-ipam\") pod \"200f01ff-d33a-474b-944e-e6ed2486db68\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.752383 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-repo-setup-combined-ca-bundle\") pod \"200f01ff-d33a-474b-944e-e6ed2486db68\" (UID: \"200f01ff-d33a-474b-944e-e6ed2486db68\") " Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.761194 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "200f01ff-d33a-474b-944e-e6ed2486db68" (UID: "200f01ff-d33a-474b-944e-e6ed2486db68"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.761211 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/200f01ff-d33a-474b-944e-e6ed2486db68-kube-api-access-6c7ph" (OuterVolumeSpecName: "kube-api-access-6c7ph") pod "200f01ff-d33a-474b-944e-e6ed2486db68" (UID: "200f01ff-d33a-474b-944e-e6ed2486db68"). InnerVolumeSpecName "kube-api-access-6c7ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.785543 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-inventory" (OuterVolumeSpecName: "inventory") pod "200f01ff-d33a-474b-944e-e6ed2486db68" (UID: "200f01ff-d33a-474b-944e-e6ed2486db68"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.785592 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "200f01ff-d33a-474b-944e-e6ed2486db68" (UID: "200f01ff-d33a-474b-944e-e6ed2486db68"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.855111 4756 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.855159 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.855171 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c7ph\" (UniqueName: \"kubernetes.io/projected/200f01ff-d33a-474b-944e-e6ed2486db68-kube-api-access-6c7ph\") on node \"crc\" DevicePath \"\"" Feb 03 09:33:22 crc kubenswrapper[4756]: I0203 09:33:22.855181 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/200f01ff-d33a-474b-944e-e6ed2486db68-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.234774 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" event={"ID":"200f01ff-d33a-474b-944e-e6ed2486db68","Type":"ContainerDied","Data":"44a7399946124b3c4a01fe9cdbf82a90d97b85587a0d8056bf60b6b30db3dd1d"} Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.234818 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44a7399946124b3c4a01fe9cdbf82a90d97b85587a0d8056bf60b6b30db3dd1d" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.234825 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.325086 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf"] Feb 03 09:33:23 crc kubenswrapper[4756]: E0203 09:33:23.326120 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="200f01ff-d33a-474b-944e-e6ed2486db68" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.326165 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="200f01ff-d33a-474b-944e-e6ed2486db68" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.326521 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="200f01ff-d33a-474b-944e-e6ed2486db68" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.327332 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.332888 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.333182 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.333398 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.342137 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.343023 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf"] Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.363686 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.363804 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbz6b\" (UniqueName: \"kubernetes.io/projected/ce26979b-e0a6-41a5-b56d-e1800d5d6704-kube-api-access-jbz6b\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.363883 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.465878 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbz6b\" (UniqueName: \"kubernetes.io/projected/ce26979b-e0a6-41a5-b56d-e1800d5d6704-kube-api-access-jbz6b\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.466226 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.466267 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.470488 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.473144 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.491341 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbz6b\" (UniqueName: \"kubernetes.io/projected/ce26979b-e0a6-41a5-b56d-e1800d5d6704-kube-api-access-jbz6b\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-xvkxf\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:23 crc kubenswrapper[4756]: I0203 09:33:23.659616 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:24 crc kubenswrapper[4756]: I0203 09:33:24.181923 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf"] Feb 03 09:33:24 crc kubenswrapper[4756]: I0203 09:33:24.245599 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" event={"ID":"ce26979b-e0a6-41a5-b56d-e1800d5d6704","Type":"ContainerStarted","Data":"d81cb170dda686f249a0a9305f7b2ee1e71b2552a175f3400c3b10847af5559d"} Feb 03 09:33:25 crc kubenswrapper[4756]: I0203 09:33:25.257046 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" event={"ID":"ce26979b-e0a6-41a5-b56d-e1800d5d6704","Type":"ContainerStarted","Data":"7d339794f25ebd184c224670d9c6b0190ed2ae85c484fad68b9f37d895d2ac25"} Feb 03 09:33:25 crc kubenswrapper[4756]: I0203 09:33:25.280981 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" podStartSLOduration=2.104733876 podStartE2EDuration="2.280962761s" podCreationTimestamp="2026-02-03 09:33:23 +0000 UTC" firstStartedPulling="2026-02-03 09:33:24.191917277 +0000 UTC m=+1395.342384652" lastFinishedPulling="2026-02-03 09:33:24.368146142 +0000 UTC m=+1395.518613537" observedRunningTime="2026-02-03 09:33:25.27396286 +0000 UTC m=+1396.424430235" watchObservedRunningTime="2026-02-03 09:33:25.280962761 +0000 UTC m=+1396.431430136" Feb 03 09:33:27 crc kubenswrapper[4756]: I0203 09:33:27.281813 4756 generic.go:334] "Generic (PLEG): container finished" podID="ce26979b-e0a6-41a5-b56d-e1800d5d6704" containerID="7d339794f25ebd184c224670d9c6b0190ed2ae85c484fad68b9f37d895d2ac25" exitCode=0 Feb 03 09:33:27 crc kubenswrapper[4756]: I0203 09:33:27.281919 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" event={"ID":"ce26979b-e0a6-41a5-b56d-e1800d5d6704","Type":"ContainerDied","Data":"7d339794f25ebd184c224670d9c6b0190ed2ae85c484fad68b9f37d895d2ac25"} Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.706145 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.766368 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbz6b\" (UniqueName: \"kubernetes.io/projected/ce26979b-e0a6-41a5-b56d-e1800d5d6704-kube-api-access-jbz6b\") pod \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.766465 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-ssh-key-openstack-edpm-ipam\") pod \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.767848 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-inventory\") pod \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\" (UID: \"ce26979b-e0a6-41a5-b56d-e1800d5d6704\") " Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.771881 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce26979b-e0a6-41a5-b56d-e1800d5d6704-kube-api-access-jbz6b" (OuterVolumeSpecName: "kube-api-access-jbz6b") pod "ce26979b-e0a6-41a5-b56d-e1800d5d6704" (UID: "ce26979b-e0a6-41a5-b56d-e1800d5d6704"). InnerVolumeSpecName "kube-api-access-jbz6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.794794 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-inventory" (OuterVolumeSpecName: "inventory") pod "ce26979b-e0a6-41a5-b56d-e1800d5d6704" (UID: "ce26979b-e0a6-41a5-b56d-e1800d5d6704"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.800213 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ce26979b-e0a6-41a5-b56d-e1800d5d6704" (UID: "ce26979b-e0a6-41a5-b56d-e1800d5d6704"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.865347 4756 scope.go:117] "RemoveContainer" containerID="3fae684069ea3d1847d0fb5eff949e9e79247538bc2dd886e0c6bf6ba83cb4a4" Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.869973 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.870006 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbz6b\" (UniqueName: \"kubernetes.io/projected/ce26979b-e0a6-41a5-b56d-e1800d5d6704-kube-api-access-jbz6b\") on node \"crc\" DevicePath \"\"" Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.870018 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce26979b-e0a6-41a5-b56d-e1800d5d6704-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:33:28 crc kubenswrapper[4756]: I0203 09:33:28.885176 4756 scope.go:117] "RemoveContainer" containerID="9fccfad52bb4f0de77a6b4eb954b6a24385b5459d1df325471e70950abb9595d" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.300048 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" event={"ID":"ce26979b-e0a6-41a5-b56d-e1800d5d6704","Type":"ContainerDied","Data":"d81cb170dda686f249a0a9305f7b2ee1e71b2552a175f3400c3b10847af5559d"} Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.300126 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d81cb170dda686f249a0a9305f7b2ee1e71b2552a175f3400c3b10847af5559d" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.300478 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-xvkxf" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.373585 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks"] Feb 03 09:33:29 crc kubenswrapper[4756]: E0203 09:33:29.373962 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce26979b-e0a6-41a5-b56d-e1800d5d6704" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.373981 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce26979b-e0a6-41a5-b56d-e1800d5d6704" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.374179 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce26979b-e0a6-41a5-b56d-e1800d5d6704" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.374761 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.380859 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.381276 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.381294 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.381912 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.419888 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks"] Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.479691 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.479758 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl64n\" (UniqueName: \"kubernetes.io/projected/5e81e9f8-2003-4c8b-8270-bf0214a074e7-kube-api-access-dl64n\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.479833 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.479908 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.581236 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.581619 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl64n\" (UniqueName: \"kubernetes.io/projected/5e81e9f8-2003-4c8b-8270-bf0214a074e7-kube-api-access-dl64n\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.581683 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.581713 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.586125 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.586503 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.586568 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.601337 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl64n\" (UniqueName: \"kubernetes.io/projected/5e81e9f8-2003-4c8b-8270-bf0214a074e7-kube-api-access-dl64n\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:29 crc kubenswrapper[4756]: I0203 09:33:29.710845 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:33:30 crc kubenswrapper[4756]: I0203 09:33:30.192318 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks"] Feb 03 09:33:30 crc kubenswrapper[4756]: I0203 09:33:30.313794 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" event={"ID":"5e81e9f8-2003-4c8b-8270-bf0214a074e7","Type":"ContainerStarted","Data":"58b41ca5683e68a8a0030cffaca0190ba0bebda8080a05075941b54b3280e265"} Feb 03 09:33:31 crc kubenswrapper[4756]: I0203 09:33:31.323615 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" event={"ID":"5e81e9f8-2003-4c8b-8270-bf0214a074e7","Type":"ContainerStarted","Data":"6b94661a334f45d1877a30c24bfd4896b9ceb4ae997e76527160b78df3afd13d"} Feb 03 09:33:31 crc kubenswrapper[4756]: I0203 09:33:31.343221 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" podStartSLOduration=2.148958265 podStartE2EDuration="2.343200105s" podCreationTimestamp="2026-02-03 09:33:29 +0000 UTC" firstStartedPulling="2026-02-03 09:33:30.197546265 +0000 UTC m=+1401.348013630" lastFinishedPulling="2026-02-03 09:33:30.391788095 +0000 UTC m=+1401.542255470" observedRunningTime="2026-02-03 09:33:31.338163277 +0000 UTC m=+1402.488630672" watchObservedRunningTime="2026-02-03 09:33:31.343200105 +0000 UTC m=+1402.493667480" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.691107 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7pgfw"] Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.693998 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.702580 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7pgfw"] Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.835341 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-catalog-content\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.835501 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-utilities\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.835536 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bknq\" (UniqueName: \"kubernetes.io/projected/19965765-4ce9-415b-bbfa-a4ee38eaa487-kube-api-access-7bknq\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.938543 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-utilities\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.938607 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bknq\" (UniqueName: \"kubernetes.io/projected/19965765-4ce9-415b-bbfa-a4ee38eaa487-kube-api-access-7bknq\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.938775 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-catalog-content\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.939219 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-catalog-content\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.939219 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-utilities\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:58 crc kubenswrapper[4756]: I0203 09:33:58.956745 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bknq\" (UniqueName: \"kubernetes.io/projected/19965765-4ce9-415b-bbfa-a4ee38eaa487-kube-api-access-7bknq\") pod \"redhat-operators-7pgfw\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:33:59 crc kubenswrapper[4756]: I0203 09:33:59.012472 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:34:00 crc kubenswrapper[4756]: I0203 09:34:00.131990 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7pgfw"] Feb 03 09:34:00 crc kubenswrapper[4756]: I0203 09:34:00.587967 4756 generic.go:334] "Generic (PLEG): container finished" podID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerID="f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7" exitCode=0 Feb 03 09:34:00 crc kubenswrapper[4756]: I0203 09:34:00.588295 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pgfw" event={"ID":"19965765-4ce9-415b-bbfa-a4ee38eaa487","Type":"ContainerDied","Data":"f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7"} Feb 03 09:34:00 crc kubenswrapper[4756]: I0203 09:34:00.588326 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pgfw" event={"ID":"19965765-4ce9-415b-bbfa-a4ee38eaa487","Type":"ContainerStarted","Data":"5f2241b1bb10a8567144374c5c6c3803387891ad7eac98bf8a26de7a7ff48806"} Feb 03 09:34:01 crc kubenswrapper[4756]: I0203 09:34:01.599930 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pgfw" event={"ID":"19965765-4ce9-415b-bbfa-a4ee38eaa487","Type":"ContainerStarted","Data":"95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04"} Feb 03 09:34:03 crc kubenswrapper[4756]: I0203 09:34:03.620295 4756 generic.go:334] "Generic (PLEG): container finished" podID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerID="95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04" exitCode=0 Feb 03 09:34:03 crc kubenswrapper[4756]: I0203 09:34:03.625791 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pgfw" event={"ID":"19965765-4ce9-415b-bbfa-a4ee38eaa487","Type":"ContainerDied","Data":"95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04"} Feb 03 09:34:04 crc kubenswrapper[4756]: I0203 09:34:04.631068 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pgfw" event={"ID":"19965765-4ce9-415b-bbfa-a4ee38eaa487","Type":"ContainerStarted","Data":"9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0"} Feb 03 09:34:04 crc kubenswrapper[4756]: I0203 09:34:04.657429 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7pgfw" podStartSLOduration=3.164893128 podStartE2EDuration="6.657411187s" podCreationTimestamp="2026-02-03 09:33:58 +0000 UTC" firstStartedPulling="2026-02-03 09:34:00.589899039 +0000 UTC m=+1431.740366434" lastFinishedPulling="2026-02-03 09:34:04.082417108 +0000 UTC m=+1435.232884493" observedRunningTime="2026-02-03 09:34:04.655884378 +0000 UTC m=+1435.806351753" watchObservedRunningTime="2026-02-03 09:34:04.657411187 +0000 UTC m=+1435.807878562" Feb 03 09:34:09 crc kubenswrapper[4756]: I0203 09:34:09.013063 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:34:09 crc kubenswrapper[4756]: I0203 09:34:09.013627 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:34:10 crc kubenswrapper[4756]: I0203 09:34:10.055684 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7pgfw" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="registry-server" probeResult="failure" output=< Feb 03 09:34:10 crc kubenswrapper[4756]: timeout: failed to connect service ":50051" within 1s Feb 03 09:34:10 crc kubenswrapper[4756]: > Feb 03 09:34:13 crc kubenswrapper[4756]: I0203 09:34:13.565969 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:34:13 crc kubenswrapper[4756]: I0203 09:34:13.566436 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:34:19 crc kubenswrapper[4756]: I0203 09:34:19.057555 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:34:19 crc kubenswrapper[4756]: I0203 09:34:19.126515 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:34:19 crc kubenswrapper[4756]: I0203 09:34:19.294590 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7pgfw"] Feb 03 09:34:20 crc kubenswrapper[4756]: I0203 09:34:20.773135 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7pgfw" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="registry-server" containerID="cri-o://9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0" gracePeriod=2 Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.200198 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.355209 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-utilities\") pod \"19965765-4ce9-415b-bbfa-a4ee38eaa487\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.355385 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-catalog-content\") pod \"19965765-4ce9-415b-bbfa-a4ee38eaa487\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.355529 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bknq\" (UniqueName: \"kubernetes.io/projected/19965765-4ce9-415b-bbfa-a4ee38eaa487-kube-api-access-7bknq\") pod \"19965765-4ce9-415b-bbfa-a4ee38eaa487\" (UID: \"19965765-4ce9-415b-bbfa-a4ee38eaa487\") " Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.356260 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-utilities" (OuterVolumeSpecName: "utilities") pod "19965765-4ce9-415b-bbfa-a4ee38eaa487" (UID: "19965765-4ce9-415b-bbfa-a4ee38eaa487"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.361970 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19965765-4ce9-415b-bbfa-a4ee38eaa487-kube-api-access-7bknq" (OuterVolumeSpecName: "kube-api-access-7bknq") pod "19965765-4ce9-415b-bbfa-a4ee38eaa487" (UID: "19965765-4ce9-415b-bbfa-a4ee38eaa487"). InnerVolumeSpecName "kube-api-access-7bknq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.458145 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bknq\" (UniqueName: \"kubernetes.io/projected/19965765-4ce9-415b-bbfa-a4ee38eaa487-kube-api-access-7bknq\") on node \"crc\" DevicePath \"\"" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.458180 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.479488 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19965765-4ce9-415b-bbfa-a4ee38eaa487" (UID: "19965765-4ce9-415b-bbfa-a4ee38eaa487"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.559599 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19965765-4ce9-415b-bbfa-a4ee38eaa487-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.785558 4756 generic.go:334] "Generic (PLEG): container finished" podID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerID="9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0" exitCode=0 Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.785674 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pgfw" event={"ID":"19965765-4ce9-415b-bbfa-a4ee38eaa487","Type":"ContainerDied","Data":"9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0"} Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.785699 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pgfw" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.786141 4756 scope.go:117] "RemoveContainer" containerID="9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.786118 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pgfw" event={"ID":"19965765-4ce9-415b-bbfa-a4ee38eaa487","Type":"ContainerDied","Data":"5f2241b1bb10a8567144374c5c6c3803387891ad7eac98bf8a26de7a7ff48806"} Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.818682 4756 scope.go:117] "RemoveContainer" containerID="95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.822431 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7pgfw"] Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.834119 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7pgfw"] Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.841625 4756 scope.go:117] "RemoveContainer" containerID="f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.883655 4756 scope.go:117] "RemoveContainer" containerID="9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0" Feb 03 09:34:21 crc kubenswrapper[4756]: E0203 09:34:21.884203 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0\": container with ID starting with 9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0 not found: ID does not exist" containerID="9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.884240 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0"} err="failed to get container status \"9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0\": rpc error: code = NotFound desc = could not find container \"9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0\": container with ID starting with 9e3fb8a9ac66eddcb7c22bbd27b967ae45bb49f777a3d238560276f9952c05a0 not found: ID does not exist" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.884269 4756 scope.go:117] "RemoveContainer" containerID="95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04" Feb 03 09:34:21 crc kubenswrapper[4756]: E0203 09:34:21.884852 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04\": container with ID starting with 95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04 not found: ID does not exist" containerID="95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.884906 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04"} err="failed to get container status \"95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04\": rpc error: code = NotFound desc = could not find container \"95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04\": container with ID starting with 95b9162dea6906c757d33f7f6a62f01558fa98a63282955768483304756cae04 not found: ID does not exist" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.884926 4756 scope.go:117] "RemoveContainer" containerID="f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7" Feb 03 09:34:21 crc kubenswrapper[4756]: E0203 09:34:21.885354 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7\": container with ID starting with f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7 not found: ID does not exist" containerID="f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7" Feb 03 09:34:21 crc kubenswrapper[4756]: I0203 09:34:21.885437 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7"} err="failed to get container status \"f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7\": rpc error: code = NotFound desc = could not find container \"f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7\": container with ID starting with f48eb40914f55cf608488cebc90db49ed6657c65f76d99eec164098a210d5cd7 not found: ID does not exist" Feb 03 09:34:23 crc kubenswrapper[4756]: I0203 09:34:23.630438 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" path="/var/lib/kubelet/pods/19965765-4ce9-415b-bbfa-a4ee38eaa487/volumes" Feb 03 09:34:29 crc kubenswrapper[4756]: I0203 09:34:29.005533 4756 scope.go:117] "RemoveContainer" containerID="84892eacef8d7ac3f04e2ee10739ab647f308ed727c7edb44f4bd3b422ee09fd" Feb 03 09:34:29 crc kubenswrapper[4756]: I0203 09:34:29.062082 4756 scope.go:117] "RemoveContainer" containerID="e2568a02083d69706f010c9cb5896146be34500a29706123dbde9047fa994f6f" Feb 03 09:34:43 crc kubenswrapper[4756]: I0203 09:34:43.566034 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:34:43 crc kubenswrapper[4756]: I0203 09:34:43.566566 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:35:13 crc kubenswrapper[4756]: I0203 09:35:13.566510 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:35:13 crc kubenswrapper[4756]: I0203 09:35:13.567046 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:35:13 crc kubenswrapper[4756]: I0203 09:35:13.567090 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:35:13 crc kubenswrapper[4756]: I0203 09:35:13.568250 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:35:13 crc kubenswrapper[4756]: I0203 09:35:13.568309 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" gracePeriod=600 Feb 03 09:35:13 crc kubenswrapper[4756]: E0203 09:35:13.689485 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:35:14 crc kubenswrapper[4756]: I0203 09:35:14.271917 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" exitCode=0 Feb 03 09:35:14 crc kubenswrapper[4756]: I0203 09:35:14.272009 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437"} Feb 03 09:35:14 crc kubenswrapper[4756]: I0203 09:35:14.272831 4756 scope.go:117] "RemoveContainer" containerID="98d1f3cc0f9ab3d41cdcdd125bb9227106916f474a34592b5e034d6f0c2ba1ac" Feb 03 09:35:14 crc kubenswrapper[4756]: I0203 09:35:14.273683 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:35:14 crc kubenswrapper[4756]: E0203 09:35:14.274044 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:35:27 crc kubenswrapper[4756]: I0203 09:35:27.614859 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:35:27 crc kubenswrapper[4756]: E0203 09:35:27.615847 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:35:29 crc kubenswrapper[4756]: I0203 09:35:29.154287 4756 scope.go:117] "RemoveContainer" containerID="f23a5200fcfb1ed2593606e2a184dd0f9adaade347b226fcad6da01d47f764b4" Feb 03 09:35:29 crc kubenswrapper[4756]: I0203 09:35:29.177640 4756 scope.go:117] "RemoveContainer" containerID="2e54bad7a03e5dc604442aba7be16afeb3ed24cbd85f34680f4cce5c88096e95" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.482860 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4t2fs"] Feb 03 09:35:40 crc kubenswrapper[4756]: E0203 09:35:40.484704 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="registry-server" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.484727 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="registry-server" Feb 03 09:35:40 crc kubenswrapper[4756]: E0203 09:35:40.484748 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="extract-content" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.484755 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="extract-content" Feb 03 09:35:40 crc kubenswrapper[4756]: E0203 09:35:40.484793 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="extract-utilities" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.484804 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="extract-utilities" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.485009 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="19965765-4ce9-415b-bbfa-a4ee38eaa487" containerName="registry-server" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.486811 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.495034 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4t2fs"] Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.687211 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-catalog-content\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.687579 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-utilities\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.687716 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvmwf\" (UniqueName: \"kubernetes.io/projected/253417ff-cd54-481f-88eb-5009ed73cc40-kube-api-access-rvmwf\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.789169 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-utilities\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.789709 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvmwf\" (UniqueName: \"kubernetes.io/projected/253417ff-cd54-481f-88eb-5009ed73cc40-kube-api-access-rvmwf\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.789707 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-utilities\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.790485 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-catalog-content\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.790867 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-catalog-content\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:40 crc kubenswrapper[4756]: I0203 09:35:40.822390 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvmwf\" (UniqueName: \"kubernetes.io/projected/253417ff-cd54-481f-88eb-5009ed73cc40-kube-api-access-rvmwf\") pod \"certified-operators-4t2fs\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:41 crc kubenswrapper[4756]: I0203 09:35:41.117731 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:41 crc kubenswrapper[4756]: I0203 09:35:41.615536 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:35:41 crc kubenswrapper[4756]: E0203 09:35:41.615965 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:35:41 crc kubenswrapper[4756]: I0203 09:35:41.626801 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4t2fs"] Feb 03 09:35:42 crc kubenswrapper[4756]: I0203 09:35:42.614927 4756 generic.go:334] "Generic (PLEG): container finished" podID="253417ff-cd54-481f-88eb-5009ed73cc40" containerID="12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806" exitCode=0 Feb 03 09:35:42 crc kubenswrapper[4756]: I0203 09:35:42.614976 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4t2fs" event={"ID":"253417ff-cd54-481f-88eb-5009ed73cc40","Type":"ContainerDied","Data":"12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806"} Feb 03 09:35:42 crc kubenswrapper[4756]: I0203 09:35:42.615011 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4t2fs" event={"ID":"253417ff-cd54-481f-88eb-5009ed73cc40","Type":"ContainerStarted","Data":"f56fdd216e0f92ed770b37ed4f425d52fd56f5a621d7247e3f41a16cd335143b"} Feb 03 09:35:43 crc kubenswrapper[4756]: I0203 09:35:43.630883 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4t2fs" event={"ID":"253417ff-cd54-481f-88eb-5009ed73cc40","Type":"ContainerStarted","Data":"87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2"} Feb 03 09:35:44 crc kubenswrapper[4756]: I0203 09:35:44.656506 4756 generic.go:334] "Generic (PLEG): container finished" podID="253417ff-cd54-481f-88eb-5009ed73cc40" containerID="87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2" exitCode=0 Feb 03 09:35:44 crc kubenswrapper[4756]: I0203 09:35:44.656667 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4t2fs" event={"ID":"253417ff-cd54-481f-88eb-5009ed73cc40","Type":"ContainerDied","Data":"87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2"} Feb 03 09:35:45 crc kubenswrapper[4756]: I0203 09:35:45.666977 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4t2fs" event={"ID":"253417ff-cd54-481f-88eb-5009ed73cc40","Type":"ContainerStarted","Data":"598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb"} Feb 03 09:35:51 crc kubenswrapper[4756]: I0203 09:35:51.119979 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:51 crc kubenswrapper[4756]: I0203 09:35:51.120377 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:51 crc kubenswrapper[4756]: I0203 09:35:51.161417 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:51 crc kubenswrapper[4756]: I0203 09:35:51.179461 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4t2fs" podStartSLOduration=8.729975371 podStartE2EDuration="11.179429491s" podCreationTimestamp="2026-02-03 09:35:40 +0000 UTC" firstStartedPulling="2026-02-03 09:35:42.616342262 +0000 UTC m=+1533.766809637" lastFinishedPulling="2026-02-03 09:35:45.065796372 +0000 UTC m=+1536.216263757" observedRunningTime="2026-02-03 09:35:45.69043516 +0000 UTC m=+1536.840902535" watchObservedRunningTime="2026-02-03 09:35:51.179429491 +0000 UTC m=+1542.329896866" Feb 03 09:35:51 crc kubenswrapper[4756]: I0203 09:35:51.759947 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:51 crc kubenswrapper[4756]: I0203 09:35:51.823190 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4t2fs"] Feb 03 09:35:53 crc kubenswrapper[4756]: I0203 09:35:53.732465 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4t2fs" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" containerName="registry-server" containerID="cri-o://598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb" gracePeriod=2 Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.235127 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.347929 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-catalog-content\") pod \"253417ff-cd54-481f-88eb-5009ed73cc40\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.348020 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvmwf\" (UniqueName: \"kubernetes.io/projected/253417ff-cd54-481f-88eb-5009ed73cc40-kube-api-access-rvmwf\") pod \"253417ff-cd54-481f-88eb-5009ed73cc40\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.348096 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-utilities\") pod \"253417ff-cd54-481f-88eb-5009ed73cc40\" (UID: \"253417ff-cd54-481f-88eb-5009ed73cc40\") " Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.349299 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-utilities" (OuterVolumeSpecName: "utilities") pod "253417ff-cd54-481f-88eb-5009ed73cc40" (UID: "253417ff-cd54-481f-88eb-5009ed73cc40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.354049 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/253417ff-cd54-481f-88eb-5009ed73cc40-kube-api-access-rvmwf" (OuterVolumeSpecName: "kube-api-access-rvmwf") pod "253417ff-cd54-481f-88eb-5009ed73cc40" (UID: "253417ff-cd54-481f-88eb-5009ed73cc40"). InnerVolumeSpecName "kube-api-access-rvmwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.449677 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.449715 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvmwf\" (UniqueName: \"kubernetes.io/projected/253417ff-cd54-481f-88eb-5009ed73cc40-kube-api-access-rvmwf\") on node \"crc\" DevicePath \"\"" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.573580 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "253417ff-cd54-481f-88eb-5009ed73cc40" (UID: "253417ff-cd54-481f-88eb-5009ed73cc40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.653156 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/253417ff-cd54-481f-88eb-5009ed73cc40-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.743515 4756 generic.go:334] "Generic (PLEG): container finished" podID="253417ff-cd54-481f-88eb-5009ed73cc40" containerID="598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb" exitCode=0 Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.743614 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4t2fs" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.743582 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4t2fs" event={"ID":"253417ff-cd54-481f-88eb-5009ed73cc40","Type":"ContainerDied","Data":"598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb"} Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.743681 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4t2fs" event={"ID":"253417ff-cd54-481f-88eb-5009ed73cc40","Type":"ContainerDied","Data":"f56fdd216e0f92ed770b37ed4f425d52fd56f5a621d7247e3f41a16cd335143b"} Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.743728 4756 scope.go:117] "RemoveContainer" containerID="598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.786128 4756 scope.go:117] "RemoveContainer" containerID="87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.789622 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4t2fs"] Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.798097 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4t2fs"] Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.806689 4756 scope.go:117] "RemoveContainer" containerID="12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.845942 4756 scope.go:117] "RemoveContainer" containerID="598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb" Feb 03 09:35:54 crc kubenswrapper[4756]: E0203 09:35:54.846595 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb\": container with ID starting with 598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb not found: ID does not exist" containerID="598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.846654 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb"} err="failed to get container status \"598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb\": rpc error: code = NotFound desc = could not find container \"598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb\": container with ID starting with 598688d0937991b0d23d0fdea303845940c30c95fa64f1bcaae7c595a70541cb not found: ID does not exist" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.846683 4756 scope.go:117] "RemoveContainer" containerID="87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2" Feb 03 09:35:54 crc kubenswrapper[4756]: E0203 09:35:54.847082 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2\": container with ID starting with 87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2 not found: ID does not exist" containerID="87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.847125 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2"} err="failed to get container status \"87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2\": rpc error: code = NotFound desc = could not find container \"87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2\": container with ID starting with 87cb701d6d343268bde64c95ec3d6c1b12239ce4daa23501c1f991507c2987a2 not found: ID does not exist" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.847158 4756 scope.go:117] "RemoveContainer" containerID="12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806" Feb 03 09:35:54 crc kubenswrapper[4756]: E0203 09:35:54.847587 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806\": container with ID starting with 12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806 not found: ID does not exist" containerID="12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806" Feb 03 09:35:54 crc kubenswrapper[4756]: I0203 09:35:54.847619 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806"} err="failed to get container status \"12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806\": rpc error: code = NotFound desc = could not find container \"12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806\": container with ID starting with 12bfc421e9b3ee72f69f2afb4ab130c69ac63dc81fc44c5186c4e69327492806 not found: ID does not exist" Feb 03 09:35:55 crc kubenswrapper[4756]: I0203 09:35:55.623537 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" path="/var/lib/kubelet/pods/253417ff-cd54-481f-88eb-5009ed73cc40/volumes" Feb 03 09:35:56 crc kubenswrapper[4756]: I0203 09:35:56.614142 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:35:56 crc kubenswrapper[4756]: E0203 09:35:56.615347 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:36:07 crc kubenswrapper[4756]: I0203 09:36:07.613761 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:36:07 crc kubenswrapper[4756]: E0203 09:36:07.615460 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:36:19 crc kubenswrapper[4756]: I0203 09:36:19.622062 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:36:19 crc kubenswrapper[4756]: E0203 09:36:19.622899 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:36:21 crc kubenswrapper[4756]: I0203 09:36:21.960672 4756 generic.go:334] "Generic (PLEG): container finished" podID="5e81e9f8-2003-4c8b-8270-bf0214a074e7" containerID="6b94661a334f45d1877a30c24bfd4896b9ceb4ae997e76527160b78df3afd13d" exitCode=0 Feb 03 09:36:21 crc kubenswrapper[4756]: I0203 09:36:21.960757 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" event={"ID":"5e81e9f8-2003-4c8b-8270-bf0214a074e7","Type":"ContainerDied","Data":"6b94661a334f45d1877a30c24bfd4896b9ceb4ae997e76527160b78df3afd13d"} Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.475223 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.614992 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-ssh-key-openstack-edpm-ipam\") pod \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.615045 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl64n\" (UniqueName: \"kubernetes.io/projected/5e81e9f8-2003-4c8b-8270-bf0214a074e7-kube-api-access-dl64n\") pod \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.615117 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-bootstrap-combined-ca-bundle\") pod \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.615197 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-inventory\") pod \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\" (UID: \"5e81e9f8-2003-4c8b-8270-bf0214a074e7\") " Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.621283 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5e81e9f8-2003-4c8b-8270-bf0214a074e7" (UID: "5e81e9f8-2003-4c8b-8270-bf0214a074e7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.623249 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e81e9f8-2003-4c8b-8270-bf0214a074e7-kube-api-access-dl64n" (OuterVolumeSpecName: "kube-api-access-dl64n") pod "5e81e9f8-2003-4c8b-8270-bf0214a074e7" (UID: "5e81e9f8-2003-4c8b-8270-bf0214a074e7"). InnerVolumeSpecName "kube-api-access-dl64n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.648201 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5e81e9f8-2003-4c8b-8270-bf0214a074e7" (UID: "5e81e9f8-2003-4c8b-8270-bf0214a074e7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.652617 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-inventory" (OuterVolumeSpecName: "inventory") pod "5e81e9f8-2003-4c8b-8270-bf0214a074e7" (UID: "5e81e9f8-2003-4c8b-8270-bf0214a074e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.717271 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.717315 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl64n\" (UniqueName: \"kubernetes.io/projected/5e81e9f8-2003-4c8b-8270-bf0214a074e7-kube-api-access-dl64n\") on node \"crc\" DevicePath \"\"" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.717328 4756 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.717341 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e81e9f8-2003-4c8b-8270-bf0214a074e7-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.986577 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" event={"ID":"5e81e9f8-2003-4c8b-8270-bf0214a074e7","Type":"ContainerDied","Data":"58b41ca5683e68a8a0030cffaca0190ba0bebda8080a05075941b54b3280e265"} Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.986959 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b41ca5683e68a8a0030cffaca0190ba0bebda8080a05075941b54b3280e265" Feb 03 09:36:23 crc kubenswrapper[4756]: I0203 09:36:23.986686 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.066769 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q"] Feb 03 09:36:24 crc kubenswrapper[4756]: E0203 09:36:24.067399 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e81e9f8-2003-4c8b-8270-bf0214a074e7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.067423 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e81e9f8-2003-4c8b-8270-bf0214a074e7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 03 09:36:24 crc kubenswrapper[4756]: E0203 09:36:24.067502 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" containerName="registry-server" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.067512 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" containerName="registry-server" Feb 03 09:36:24 crc kubenswrapper[4756]: E0203 09:36:24.067525 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" containerName="extract-utilities" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.067534 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" containerName="extract-utilities" Feb 03 09:36:24 crc kubenswrapper[4756]: E0203 09:36:24.067558 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" containerName="extract-content" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.067566 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" containerName="extract-content" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.067807 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e81e9f8-2003-4c8b-8270-bf0214a074e7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.067830 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="253417ff-cd54-481f-88eb-5009ed73cc40" containerName="registry-server" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.068878 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.070938 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.071156 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.071229 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.071339 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.080078 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q"] Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.225658 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.225860 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.225925 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bj76\" (UniqueName: \"kubernetes.io/projected/2847ae70-333c-420d-b7a6-2c547477f35b-kube-api-access-5bj76\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.327643 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.327786 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.327834 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bj76\" (UniqueName: \"kubernetes.io/projected/2847ae70-333c-420d-b7a6-2c547477f35b-kube-api-access-5bj76\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.331409 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.331436 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.353695 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bj76\" (UniqueName: \"kubernetes.io/projected/2847ae70-333c-420d-b7a6-2c547477f35b-kube-api-access-5bj76\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-blt9q\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.385211 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.908129 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q"] Feb 03 09:36:24 crc kubenswrapper[4756]: I0203 09:36:24.996976 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" event={"ID":"2847ae70-333c-420d-b7a6-2c547477f35b","Type":"ContainerStarted","Data":"fe33b6ea9aa0c00f4b8b349ac86c83a371ba696c2ae279df944b308572a850a3"} Feb 03 09:36:26 crc kubenswrapper[4756]: I0203 09:36:26.006185 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" event={"ID":"2847ae70-333c-420d-b7a6-2c547477f35b","Type":"ContainerStarted","Data":"a7ff8d44f0a8ad32784e6a6f0d76aff74df474a2d6303738e37ce07ddfaf6f47"} Feb 03 09:36:26 crc kubenswrapper[4756]: I0203 09:36:26.036512 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" podStartSLOduration=1.863866417 podStartE2EDuration="2.036491898s" podCreationTimestamp="2026-02-03 09:36:24 +0000 UTC" firstStartedPulling="2026-02-03 09:36:24.91321441 +0000 UTC m=+1576.063681775" lastFinishedPulling="2026-02-03 09:36:25.085839881 +0000 UTC m=+1576.236307256" observedRunningTime="2026-02-03 09:36:26.025321437 +0000 UTC m=+1577.175788832" watchObservedRunningTime="2026-02-03 09:36:26.036491898 +0000 UTC m=+1577.186959273" Feb 03 09:36:29 crc kubenswrapper[4756]: I0203 09:36:29.210962 4756 scope.go:117] "RemoveContainer" containerID="89e30e6e3cbb4c978aca36fed12ee6edf3990818c07bf3f57973f1831ed38f26" Feb 03 09:36:29 crc kubenswrapper[4756]: I0203 09:36:29.233901 4756 scope.go:117] "RemoveContainer" containerID="18a1fa4d8a0e19c6ce4bb001a39b44f449d70c3d413770a3888e3fd9a4846b73" Feb 03 09:36:29 crc kubenswrapper[4756]: I0203 09:36:29.255924 4756 scope.go:117] "RemoveContainer" containerID="2fe920db5b7ff2901fe8a303936aaa9ddb30bd04c5c80cf24c7208d76d56de96" Feb 03 09:36:32 crc kubenswrapper[4756]: I0203 09:36:32.614948 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:36:32 crc kubenswrapper[4756]: E0203 09:36:32.615806 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.389496 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sr5th"] Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.393269 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.402597 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sr5th"] Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.471342 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng5fk\" (UniqueName: \"kubernetes.io/projected/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-kube-api-access-ng5fk\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.471391 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-catalog-content\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.471571 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-utilities\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.572600 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng5fk\" (UniqueName: \"kubernetes.io/projected/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-kube-api-access-ng5fk\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.572646 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-catalog-content\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.572795 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-utilities\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.573342 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-utilities\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.573375 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-catalog-content\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.591536 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng5fk\" (UniqueName: \"kubernetes.io/projected/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-kube-api-access-ng5fk\") pod \"community-operators-sr5th\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:41 crc kubenswrapper[4756]: I0203 09:36:41.724928 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:42 crc kubenswrapper[4756]: I0203 09:36:42.251633 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sr5th"] Feb 03 09:36:43 crc kubenswrapper[4756]: I0203 09:36:43.140941 4756 generic.go:334] "Generic (PLEG): container finished" podID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerID="854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab" exitCode=0 Feb 03 09:36:43 crc kubenswrapper[4756]: I0203 09:36:43.141048 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sr5th" event={"ID":"dc5b0929-b05c-4590-bb99-08f0c9aea9ee","Type":"ContainerDied","Data":"854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab"} Feb 03 09:36:43 crc kubenswrapper[4756]: I0203 09:36:43.141253 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sr5th" event={"ID":"dc5b0929-b05c-4590-bb99-08f0c9aea9ee","Type":"ContainerStarted","Data":"1059de0abcd5cb0927a64ff29329f23efe8a6c59ca2ddeff8b2f792165797c07"} Feb 03 09:36:45 crc kubenswrapper[4756]: I0203 09:36:45.164121 4756 generic.go:334] "Generic (PLEG): container finished" podID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerID="44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065" exitCode=0 Feb 03 09:36:45 crc kubenswrapper[4756]: I0203 09:36:45.164224 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sr5th" event={"ID":"dc5b0929-b05c-4590-bb99-08f0c9aea9ee","Type":"ContainerDied","Data":"44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065"} Feb 03 09:36:46 crc kubenswrapper[4756]: I0203 09:36:46.178215 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sr5th" event={"ID":"dc5b0929-b05c-4590-bb99-08f0c9aea9ee","Type":"ContainerStarted","Data":"7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc"} Feb 03 09:36:46 crc kubenswrapper[4756]: I0203 09:36:46.201418 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sr5th" podStartSLOduration=2.527193263 podStartE2EDuration="5.201397062s" podCreationTimestamp="2026-02-03 09:36:41 +0000 UTC" firstStartedPulling="2026-02-03 09:36:43.143199914 +0000 UTC m=+1594.293667289" lastFinishedPulling="2026-02-03 09:36:45.817403683 +0000 UTC m=+1596.967871088" observedRunningTime="2026-02-03 09:36:46.193152653 +0000 UTC m=+1597.343620028" watchObservedRunningTime="2026-02-03 09:36:46.201397062 +0000 UTC m=+1597.351864437" Feb 03 09:36:47 crc kubenswrapper[4756]: I0203 09:36:47.615204 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:36:47 crc kubenswrapper[4756]: E0203 09:36:47.615668 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:36:51 crc kubenswrapper[4756]: I0203 09:36:51.726380 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:51 crc kubenswrapper[4756]: I0203 09:36:51.726799 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:51 crc kubenswrapper[4756]: I0203 09:36:51.787414 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:52 crc kubenswrapper[4756]: I0203 09:36:52.278707 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:52 crc kubenswrapper[4756]: I0203 09:36:52.324356 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sr5th"] Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.247158 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sr5th" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerName="registry-server" containerID="cri-o://7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc" gracePeriod=2 Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.714569 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.833566 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-utilities\") pod \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.833768 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-catalog-content\") pod \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.833879 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng5fk\" (UniqueName: \"kubernetes.io/projected/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-kube-api-access-ng5fk\") pod \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\" (UID: \"dc5b0929-b05c-4590-bb99-08f0c9aea9ee\") " Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.834502 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-utilities" (OuterVolumeSpecName: "utilities") pod "dc5b0929-b05c-4590-bb99-08f0c9aea9ee" (UID: "dc5b0929-b05c-4590-bb99-08f0c9aea9ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.839297 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-kube-api-access-ng5fk" (OuterVolumeSpecName: "kube-api-access-ng5fk") pod "dc5b0929-b05c-4590-bb99-08f0c9aea9ee" (UID: "dc5b0929-b05c-4590-bb99-08f0c9aea9ee"). InnerVolumeSpecName "kube-api-access-ng5fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.888418 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc5b0929-b05c-4590-bb99-08f0c9aea9ee" (UID: "dc5b0929-b05c-4590-bb99-08f0c9aea9ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.935907 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.935947 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng5fk\" (UniqueName: \"kubernetes.io/projected/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-kube-api-access-ng5fk\") on node \"crc\" DevicePath \"\"" Feb 03 09:36:54 crc kubenswrapper[4756]: I0203 09:36:54.935959 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5b0929-b05c-4590-bb99-08f0c9aea9ee-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.285353 4756 generic.go:334] "Generic (PLEG): container finished" podID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerID="7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc" exitCode=0 Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.285404 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sr5th" event={"ID":"dc5b0929-b05c-4590-bb99-08f0c9aea9ee","Type":"ContainerDied","Data":"7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc"} Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.285424 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sr5th" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.285455 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sr5th" event={"ID":"dc5b0929-b05c-4590-bb99-08f0c9aea9ee","Type":"ContainerDied","Data":"1059de0abcd5cb0927a64ff29329f23efe8a6c59ca2ddeff8b2f792165797c07"} Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.285477 4756 scope.go:117] "RemoveContainer" containerID="7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.323296 4756 scope.go:117] "RemoveContainer" containerID="44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.333939 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sr5th"] Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.342469 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sr5th"] Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.356020 4756 scope.go:117] "RemoveContainer" containerID="854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.402405 4756 scope.go:117] "RemoveContainer" containerID="7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc" Feb 03 09:36:55 crc kubenswrapper[4756]: E0203 09:36:55.403001 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc\": container with ID starting with 7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc not found: ID does not exist" containerID="7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.403046 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc"} err="failed to get container status \"7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc\": rpc error: code = NotFound desc = could not find container \"7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc\": container with ID starting with 7abb19841f2cc4379d67ba82a4455922ed54af560adda169e1aba0f8113eeefc not found: ID does not exist" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.403072 4756 scope.go:117] "RemoveContainer" containerID="44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065" Feb 03 09:36:55 crc kubenswrapper[4756]: E0203 09:36:55.403480 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065\": container with ID starting with 44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065 not found: ID does not exist" containerID="44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.403576 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065"} err="failed to get container status \"44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065\": rpc error: code = NotFound desc = could not find container \"44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065\": container with ID starting with 44d6697a249e95476323ba5821f99f0dcfb5aa67bbb89938459fb99c6c7d7065 not found: ID does not exist" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.403642 4756 scope.go:117] "RemoveContainer" containerID="854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab" Feb 03 09:36:55 crc kubenswrapper[4756]: E0203 09:36:55.403915 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab\": container with ID starting with 854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab not found: ID does not exist" containerID="854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.403940 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab"} err="failed to get container status \"854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab\": rpc error: code = NotFound desc = could not find container \"854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab\": container with ID starting with 854a84958bfeb84d524b52eb2b506fc03b89a6b20a22f940dc834323dd2386ab not found: ID does not exist" Feb 03 09:36:55 crc kubenswrapper[4756]: I0203 09:36:55.627550 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" path="/var/lib/kubelet/pods/dc5b0929-b05c-4590-bb99-08f0c9aea9ee/volumes" Feb 03 09:36:58 crc kubenswrapper[4756]: I0203 09:36:58.614727 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:36:58 crc kubenswrapper[4756]: E0203 09:36:58.615247 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:37:10 crc kubenswrapper[4756]: I0203 09:37:10.053886 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-sl9vg"] Feb 03 09:37:10 crc kubenswrapper[4756]: I0203 09:37:10.062719 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-aaef-account-create-update-cjsjc"] Feb 03 09:37:10 crc kubenswrapper[4756]: I0203 09:37:10.071887 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-187d-account-create-update-8rkgm"] Feb 03 09:37:10 crc kubenswrapper[4756]: I0203 09:37:10.080482 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-sl9vg"] Feb 03 09:37:10 crc kubenswrapper[4756]: I0203 09:37:10.087493 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-187d-account-create-update-8rkgm"] Feb 03 09:37:10 crc kubenswrapper[4756]: I0203 09:37:10.095278 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-aaef-account-create-update-cjsjc"] Feb 03 09:37:10 crc kubenswrapper[4756]: I0203 09:37:10.614060 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:37:10 crc kubenswrapper[4756]: E0203 09:37:10.614586 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:37:11 crc kubenswrapper[4756]: I0203 09:37:11.027883 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-lzqmn"] Feb 03 09:37:11 crc kubenswrapper[4756]: I0203 09:37:11.037278 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-lzqmn"] Feb 03 09:37:11 crc kubenswrapper[4756]: I0203 09:37:11.637231 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="344e964e-c1ec-4fdc-82a3-223389b11f1a" path="/var/lib/kubelet/pods/344e964e-c1ec-4fdc-82a3-223389b11f1a/volumes" Feb 03 09:37:11 crc kubenswrapper[4756]: I0203 09:37:11.640050 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d05d9f3-e465-481b-92e0-c4d2c8c640bb" path="/var/lib/kubelet/pods/4d05d9f3-e465-481b-92e0-c4d2c8c640bb/volumes" Feb 03 09:37:11 crc kubenswrapper[4756]: I0203 09:37:11.642363 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d909b510-1424-40ce-ad72-0b46a5304b42" path="/var/lib/kubelet/pods/d909b510-1424-40ce-ad72-0b46a5304b42/volumes" Feb 03 09:37:11 crc kubenswrapper[4756]: I0203 09:37:11.643618 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc25ed95-beda-41da-b65f-10dea4ee003f" path="/var/lib/kubelet/pods/dc25ed95-beda-41da-b65f-10dea4ee003f/volumes" Feb 03 09:37:14 crc kubenswrapper[4756]: I0203 09:37:14.035588 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-kdsms"] Feb 03 09:37:14 crc kubenswrapper[4756]: I0203 09:37:14.049592 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-60b6-account-create-update-xwbqp"] Feb 03 09:37:14 crc kubenswrapper[4756]: I0203 09:37:14.059785 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-60b6-account-create-update-xwbqp"] Feb 03 09:37:14 crc kubenswrapper[4756]: I0203 09:37:14.070050 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-kdsms"] Feb 03 09:37:15 crc kubenswrapper[4756]: I0203 09:37:15.624032 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a643eeaf-1782-49f9-bbb7-55bdd019f68e" path="/var/lib/kubelet/pods/a643eeaf-1782-49f9-bbb7-55bdd019f68e/volumes" Feb 03 09:37:15 crc kubenswrapper[4756]: I0203 09:37:15.625158 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0729c9f-8d1d-49a8-9cff-eefbcafac7b0" path="/var/lib/kubelet/pods/f0729c9f-8d1d-49a8-9cff-eefbcafac7b0/volumes" Feb 03 09:37:23 crc kubenswrapper[4756]: I0203 09:37:23.614723 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:37:23 crc kubenswrapper[4756]: E0203 09:37:23.615272 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:37:26 crc kubenswrapper[4756]: I0203 09:37:26.031510 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-bcd5t"] Feb 03 09:37:26 crc kubenswrapper[4756]: I0203 09:37:26.041366 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-bcd5t"] Feb 03 09:37:27 crc kubenswrapper[4756]: I0203 09:37:27.625346 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa4bceb9-3d5a-4fe4-9b98-ed408d36116b" path="/var/lib/kubelet/pods/aa4bceb9-3d5a-4fe4-9b98-ed408d36116b/volumes" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.341289 4756 scope.go:117] "RemoveContainer" containerID="e94495bb22c6c2245ab14c6e8c70298201b7820743e32866bbf019e7ba9f0112" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.376600 4756 scope.go:117] "RemoveContainer" containerID="3c157f626593654ad5af135ca36c79030fd04fed353f1a93b0fac268ec847efb" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.428293 4756 scope.go:117] "RemoveContainer" containerID="c71513e6b7ef7e9c0014b955513855c12ac490eb1d6cf98839ffd582b23fd90d" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.451400 4756 scope.go:117] "RemoveContainer" containerID="478585637a4c698a0a8007b3105b7f26fdca2a838eb61dce5934385134095b6b" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.501287 4756 scope.go:117] "RemoveContainer" containerID="bbcbd081925ae74379917495609fb1d18e3156dbcee7f2b3c972d0a9d38cc618" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.517736 4756 scope.go:117] "RemoveContainer" containerID="9d50aed8cb96d489fc52415db103e11db5ee939cc96736b7234b76e4adfa48ef" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.536552 4756 scope.go:117] "RemoveContainer" containerID="de0dcf48427ffdf38b3cb5ad420617b2892e5f15c91ff1d65746372eff9f8f41" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.583628 4756 scope.go:117] "RemoveContainer" containerID="13ba6bda4e41fd693ffb839cd47a69928c65ae3146960eca54c63fe08033b092" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.637215 4756 scope.go:117] "RemoveContainer" containerID="936d4ddfbad84d71090688ba2d7ea6e1d5e4bfc558f243446d08a9d6339686a1" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.655547 4756 scope.go:117] "RemoveContainer" containerID="01edc87ebba40d6af61ecd48708d7c0f128ea4ffc8b8e29e0467a010a577e84e" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.673703 4756 scope.go:117] "RemoveContainer" containerID="891ceadb7fb03f03a5e991cb3032159128fc9153668570a7cf316d803966f9e1" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.691941 4756 scope.go:117] "RemoveContainer" containerID="6c943329f74b1bdcce52b7e8805fa9b7f41ab6f68b547aac834cb754b75b62c6" Feb 03 09:37:29 crc kubenswrapper[4756]: I0203 09:37:29.751148 4756 scope.go:117] "RemoveContainer" containerID="0c4c1352d1710a022c86866da9b6b5f4cae10388c10fde9d55f34c5985e208ec" Feb 03 09:37:35 crc kubenswrapper[4756]: I0203 09:37:35.615031 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:37:35 crc kubenswrapper[4756]: E0203 09:37:35.616065 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:37:42 crc kubenswrapper[4756]: I0203 09:37:42.741314 4756 generic.go:334] "Generic (PLEG): container finished" podID="2847ae70-333c-420d-b7a6-2c547477f35b" containerID="a7ff8d44f0a8ad32784e6a6f0d76aff74df474a2d6303738e37ce07ddfaf6f47" exitCode=0 Feb 03 09:37:42 crc kubenswrapper[4756]: I0203 09:37:42.741504 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" event={"ID":"2847ae70-333c-420d-b7a6-2c547477f35b","Type":"ContainerDied","Data":"a7ff8d44f0a8ad32784e6a6f0d76aff74df474a2d6303738e37ce07ddfaf6f47"} Feb 03 09:37:43 crc kubenswrapper[4756]: I0203 09:37:43.048178 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-8j6l5"] Feb 03 09:37:43 crc kubenswrapper[4756]: I0203 09:37:43.057985 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-8j6l5"] Feb 03 09:37:43 crc kubenswrapper[4756]: I0203 09:37:43.632039 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9" path="/var/lib/kubelet/pods/d3a4b6f6-ed23-4733-a1e8-eaf7e898ffe9/volumes" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.172617 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.305945 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-inventory\") pod \"2847ae70-333c-420d-b7a6-2c547477f35b\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.306027 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-ssh-key-openstack-edpm-ipam\") pod \"2847ae70-333c-420d-b7a6-2c547477f35b\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.306105 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bj76\" (UniqueName: \"kubernetes.io/projected/2847ae70-333c-420d-b7a6-2c547477f35b-kube-api-access-5bj76\") pod \"2847ae70-333c-420d-b7a6-2c547477f35b\" (UID: \"2847ae70-333c-420d-b7a6-2c547477f35b\") " Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.314993 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2847ae70-333c-420d-b7a6-2c547477f35b-kube-api-access-5bj76" (OuterVolumeSpecName: "kube-api-access-5bj76") pod "2847ae70-333c-420d-b7a6-2c547477f35b" (UID: "2847ae70-333c-420d-b7a6-2c547477f35b"). InnerVolumeSpecName "kube-api-access-5bj76". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.331908 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-inventory" (OuterVolumeSpecName: "inventory") pod "2847ae70-333c-420d-b7a6-2c547477f35b" (UID: "2847ae70-333c-420d-b7a6-2c547477f35b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.335328 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2847ae70-333c-420d-b7a6-2c547477f35b" (UID: "2847ae70-333c-420d-b7a6-2c547477f35b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.408327 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.408365 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2847ae70-333c-420d-b7a6-2c547477f35b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.408381 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bj76\" (UniqueName: \"kubernetes.io/projected/2847ae70-333c-420d-b7a6-2c547477f35b-kube-api-access-5bj76\") on node \"crc\" DevicePath \"\"" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.766962 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" event={"ID":"2847ae70-333c-420d-b7a6-2c547477f35b","Type":"ContainerDied","Data":"fe33b6ea9aa0c00f4b8b349ac86c83a371ba696c2ae279df944b308572a850a3"} Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.767001 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe33b6ea9aa0c00f4b8b349ac86c83a371ba696c2ae279df944b308572a850a3" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.767051 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-blt9q" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.867369 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2"] Feb 03 09:37:44 crc kubenswrapper[4756]: E0203 09:37:44.867762 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2847ae70-333c-420d-b7a6-2c547477f35b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.867779 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="2847ae70-333c-420d-b7a6-2c547477f35b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 03 09:37:44 crc kubenswrapper[4756]: E0203 09:37:44.867787 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerName="extract-content" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.867793 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerName="extract-content" Feb 03 09:37:44 crc kubenswrapper[4756]: E0203 09:37:44.867815 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerName="extract-utilities" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.867822 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerName="extract-utilities" Feb 03 09:37:44 crc kubenswrapper[4756]: E0203 09:37:44.867838 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerName="registry-server" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.867843 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerName="registry-server" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.868037 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5b0929-b05c-4590-bb99-08f0c9aea9ee" containerName="registry-server" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.868056 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="2847ae70-333c-420d-b7a6-2c547477f35b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.868795 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.871031 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.871414 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.871575 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.872352 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:37:44 crc kubenswrapper[4756]: I0203 09:37:44.882216 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2"] Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.019936 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.020097 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.020159 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnmnv\" (UniqueName: \"kubernetes.io/projected/3c14e04f-9c29-45ae-84dc-bad543ff55d2-kube-api-access-dnmnv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.121781 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.121902 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.121961 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnmnv\" (UniqueName: \"kubernetes.io/projected/3c14e04f-9c29-45ae-84dc-bad543ff55d2-kube-api-access-dnmnv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.126827 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.131930 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.139590 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnmnv\" (UniqueName: \"kubernetes.io/projected/3c14e04f-9c29-45ae-84dc-bad543ff55d2-kube-api-access-dnmnv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.185592 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.701226 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2"] Feb 03 09:37:45 crc kubenswrapper[4756]: I0203 09:37:45.778481 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" event={"ID":"3c14e04f-9c29-45ae-84dc-bad543ff55d2","Type":"ContainerStarted","Data":"35bb1967f23b490db391aed68589c6942ff349a028becb23af2c2e31bf4011b1"} Feb 03 09:37:46 crc kubenswrapper[4756]: I0203 09:37:46.788123 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" event={"ID":"3c14e04f-9c29-45ae-84dc-bad543ff55d2","Type":"ContainerStarted","Data":"56133ba6daf6f471702b88ee8b5f84d58f1a6c7f09f8a21ff8dadda1fa13bf40"} Feb 03 09:37:46 crc kubenswrapper[4756]: I0203 09:37:46.812873 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" podStartSLOduration=2.636582187 podStartE2EDuration="2.812850326s" podCreationTimestamp="2026-02-03 09:37:44 +0000 UTC" firstStartedPulling="2026-02-03 09:37:45.706806579 +0000 UTC m=+1656.857273954" lastFinishedPulling="2026-02-03 09:37:45.883074698 +0000 UTC m=+1657.033542093" observedRunningTime="2026-02-03 09:37:46.807281832 +0000 UTC m=+1657.957749227" watchObservedRunningTime="2026-02-03 09:37:46.812850326 +0000 UTC m=+1657.963317721" Feb 03 09:37:48 crc kubenswrapper[4756]: I0203 09:37:48.614655 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:37:48 crc kubenswrapper[4756]: E0203 09:37:48.615343 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.033555 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-nz7sg"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.045023 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-88aa-account-create-update-ndhz5"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.060528 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-nz7sg"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.077238 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-88aa-account-create-update-ndhz5"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.089384 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-hgsfq"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.099859 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-hgsfq"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.111009 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-953f-account-create-update-595f4"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.144526 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-b948-account-create-update-q4k64"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.152814 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-953f-account-create-update-595f4"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.161648 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-b948-account-create-update-q4k64"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.170007 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-jbrwj"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.176970 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-jbrwj"] Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.626393 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f0bf352-90f0-4613-b6c8-9206ddbc4235" path="/var/lib/kubelet/pods/0f0bf352-90f0-4613-b6c8-9206ddbc4235/volumes" Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.627181 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5971a24c-30d3-4e3e-a649-6d6f637dbfd5" path="/var/lib/kubelet/pods/5971a24c-30d3-4e3e-a649-6d6f637dbfd5/volumes" Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.627753 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e555442-e24a-467c-8ab3-426856e59e43" path="/var/lib/kubelet/pods/7e555442-e24a-467c-8ab3-426856e59e43/volumes" Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.628246 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4e6768f-7cd7-4a37-ab83-59b8fa562c11" path="/var/lib/kubelet/pods/a4e6768f-7cd7-4a37-ab83-59b8fa562c11/volumes" Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.629228 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad7fc552-1c39-4c32-a9ea-7df1556ce73f" path="/var/lib/kubelet/pods/ad7fc552-1c39-4c32-a9ea-7df1556ce73f/volumes" Feb 03 09:37:51 crc kubenswrapper[4756]: I0203 09:37:51.629739 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051" path="/var/lib/kubelet/pods/e4f8b88f-ba51-4ac5-b046-4bcb8e0b4051/volumes" Feb 03 09:37:58 crc kubenswrapper[4756]: I0203 09:37:58.028092 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-h5vn4"] Feb 03 09:37:58 crc kubenswrapper[4756]: I0203 09:37:58.037950 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-h5vn4"] Feb 03 09:37:59 crc kubenswrapper[4756]: I0203 09:37:59.624732 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e62dcdb-516f-406c-a6dd-d780f6c250bc" path="/var/lib/kubelet/pods/3e62dcdb-516f-406c-a6dd-d780f6c250bc/volumes" Feb 03 09:38:00 crc kubenswrapper[4756]: I0203 09:38:00.614226 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:38:00 crc kubenswrapper[4756]: E0203 09:38:00.614616 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:38:15 crc kubenswrapper[4756]: I0203 09:38:15.614871 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:38:15 crc kubenswrapper[4756]: E0203 09:38:15.616461 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.193778 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gz5td"] Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.196761 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.205019 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gz5td"] Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.289933 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh4wn\" (UniqueName: \"kubernetes.io/projected/31805a74-ec1e-4f9d-806d-5e77fe88f709-kube-api-access-fh4wn\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.290035 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-catalog-content\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.290112 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-utilities\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.391882 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh4wn\" (UniqueName: \"kubernetes.io/projected/31805a74-ec1e-4f9d-806d-5e77fe88f709-kube-api-access-fh4wn\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.391968 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-catalog-content\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.392023 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-utilities\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.392777 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-utilities\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.392883 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-catalog-content\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.410831 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh4wn\" (UniqueName: \"kubernetes.io/projected/31805a74-ec1e-4f9d-806d-5e77fe88f709-kube-api-access-fh4wn\") pod \"redhat-marketplace-gz5td\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:19 crc kubenswrapper[4756]: I0203 09:38:19.530344 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:20 crc kubenswrapper[4756]: I0203 09:38:20.037682 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gz5td"] Feb 03 09:38:20 crc kubenswrapper[4756]: I0203 09:38:20.075907 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gz5td" event={"ID":"31805a74-ec1e-4f9d-806d-5e77fe88f709","Type":"ContainerStarted","Data":"be448f7e5db749d4ba0d1907f1ddd959b877a0339a86a40cef594cab7fbc5c3a"} Feb 03 09:38:21 crc kubenswrapper[4756]: I0203 09:38:21.086418 4756 generic.go:334] "Generic (PLEG): container finished" podID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerID="9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6" exitCode=0 Feb 03 09:38:21 crc kubenswrapper[4756]: I0203 09:38:21.086513 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gz5td" event={"ID":"31805a74-ec1e-4f9d-806d-5e77fe88f709","Type":"ContainerDied","Data":"9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6"} Feb 03 09:38:21 crc kubenswrapper[4756]: I0203 09:38:21.089341 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 09:38:22 crc kubenswrapper[4756]: I0203 09:38:22.095853 4756 generic.go:334] "Generic (PLEG): container finished" podID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerID="9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f" exitCode=0 Feb 03 09:38:22 crc kubenswrapper[4756]: I0203 09:38:22.095952 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gz5td" event={"ID":"31805a74-ec1e-4f9d-806d-5e77fe88f709","Type":"ContainerDied","Data":"9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f"} Feb 03 09:38:23 crc kubenswrapper[4756]: I0203 09:38:23.106284 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gz5td" event={"ID":"31805a74-ec1e-4f9d-806d-5e77fe88f709","Type":"ContainerStarted","Data":"472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9"} Feb 03 09:38:23 crc kubenswrapper[4756]: I0203 09:38:23.135823 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gz5td" podStartSLOduration=2.7092458219999997 podStartE2EDuration="4.135798398s" podCreationTimestamp="2026-02-03 09:38:19 +0000 UTC" firstStartedPulling="2026-02-03 09:38:21.088780463 +0000 UTC m=+1692.239247838" lastFinishedPulling="2026-02-03 09:38:22.515333019 +0000 UTC m=+1693.665800414" observedRunningTime="2026-02-03 09:38:23.123593047 +0000 UTC m=+1694.274060432" watchObservedRunningTime="2026-02-03 09:38:23.135798398 +0000 UTC m=+1694.286265773" Feb 03 09:38:26 crc kubenswrapper[4756]: I0203 09:38:26.047029 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4llhn"] Feb 03 09:38:26 crc kubenswrapper[4756]: I0203 09:38:26.055603 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4llhn"] Feb 03 09:38:27 crc kubenswrapper[4756]: I0203 09:38:27.628571 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6f14305-2e15-4ccc-ae43-230f1f8daf18" path="/var/lib/kubelet/pods/f6f14305-2e15-4ccc-ae43-230f1f8daf18/volumes" Feb 03 09:38:29 crc kubenswrapper[4756]: I0203 09:38:29.531241 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:29 crc kubenswrapper[4756]: I0203 09:38:29.531509 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:29 crc kubenswrapper[4756]: I0203 09:38:29.578191 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:29 crc kubenswrapper[4756]: I0203 09:38:29.918321 4756 scope.go:117] "RemoveContainer" containerID="57832a58c9733cca830b6ae89f7f260772e41d1643546ecd029a943252c27533" Feb 03 09:38:29 crc kubenswrapper[4756]: I0203 09:38:29.968028 4756 scope.go:117] "RemoveContainer" containerID="14aca9176edc225e01b661d9c251b5f3be6bd3785c25e8068a9c1eec85d47e16" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.009968 4756 scope.go:117] "RemoveContainer" containerID="3dddb67456e4c4f60bdd8c77ab68cab576f8d923ec669b7642248cc8b0cc0057" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.052845 4756 scope.go:117] "RemoveContainer" containerID="ae612df2dc0b954d698378e015dc7145ec044a37df2bf55abf1e2455834b5982" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.091220 4756 scope.go:117] "RemoveContainer" containerID="896fafbfcde87e7c0add95627b1ce3d650a8c3ebfaf4f765df460f2347945fbd" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.150584 4756 scope.go:117] "RemoveContainer" containerID="ab54b30466202333f111da6ab5eaae031f752ead2870986ae79ef42231ddb2ed" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.181001 4756 scope.go:117] "RemoveContainer" containerID="8ca8f574bcef2d1e934268f84949aeeeba5ee177f6c8ef940f7cbd5ca210ee30" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.227212 4756 scope.go:117] "RemoveContainer" containerID="af49ff2cd7d568d6589b3743ca0d88d0f53bcb4d7a4d0881699757d727ee9212" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.259465 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.311045 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gz5td"] Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.327483 4756 scope.go:117] "RemoveContainer" containerID="4dec43cbd7de72379ea3fe3da06372a55dceb25a47b97cfb6a9f05613c52507e" Feb 03 09:38:30 crc kubenswrapper[4756]: I0203 09:38:30.614056 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:38:30 crc kubenswrapper[4756]: E0203 09:38:30.614308 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:38:32 crc kubenswrapper[4756]: I0203 09:38:32.218634 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gz5td" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerName="registry-server" containerID="cri-o://472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9" gracePeriod=2 Feb 03 09:38:32 crc kubenswrapper[4756]: I0203 09:38:32.997819 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.044972 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh4wn\" (UniqueName: \"kubernetes.io/projected/31805a74-ec1e-4f9d-806d-5e77fe88f709-kube-api-access-fh4wn\") pod \"31805a74-ec1e-4f9d-806d-5e77fe88f709\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.045198 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-catalog-content\") pod \"31805a74-ec1e-4f9d-806d-5e77fe88f709\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.045246 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-utilities\") pod \"31805a74-ec1e-4f9d-806d-5e77fe88f709\" (UID: \"31805a74-ec1e-4f9d-806d-5e77fe88f709\") " Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.047287 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-utilities" (OuterVolumeSpecName: "utilities") pod "31805a74-ec1e-4f9d-806d-5e77fe88f709" (UID: "31805a74-ec1e-4f9d-806d-5e77fe88f709"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.062744 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31805a74-ec1e-4f9d-806d-5e77fe88f709-kube-api-access-fh4wn" (OuterVolumeSpecName: "kube-api-access-fh4wn") pod "31805a74-ec1e-4f9d-806d-5e77fe88f709" (UID: "31805a74-ec1e-4f9d-806d-5e77fe88f709"). InnerVolumeSpecName "kube-api-access-fh4wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.081580 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31805a74-ec1e-4f9d-806d-5e77fe88f709" (UID: "31805a74-ec1e-4f9d-806d-5e77fe88f709"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.148459 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.148499 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31805a74-ec1e-4f9d-806d-5e77fe88f709-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.148512 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh4wn\" (UniqueName: \"kubernetes.io/projected/31805a74-ec1e-4f9d-806d-5e77fe88f709-kube-api-access-fh4wn\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.231543 4756 generic.go:334] "Generic (PLEG): container finished" podID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerID="472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9" exitCode=0 Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.231587 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gz5td" event={"ID":"31805a74-ec1e-4f9d-806d-5e77fe88f709","Type":"ContainerDied","Data":"472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9"} Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.231611 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gz5td" event={"ID":"31805a74-ec1e-4f9d-806d-5e77fe88f709","Type":"ContainerDied","Data":"be448f7e5db749d4ba0d1907f1ddd959b877a0339a86a40cef594cab7fbc5c3a"} Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.231626 4756 scope.go:117] "RemoveContainer" containerID="472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.231735 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gz5td" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.257818 4756 scope.go:117] "RemoveContainer" containerID="9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.263761 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gz5td"] Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.289211 4756 scope.go:117] "RemoveContainer" containerID="9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.293614 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gz5td"] Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.330512 4756 scope.go:117] "RemoveContainer" containerID="472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9" Feb 03 09:38:33 crc kubenswrapper[4756]: E0203 09:38:33.331668 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9\": container with ID starting with 472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9 not found: ID does not exist" containerID="472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.331729 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9"} err="failed to get container status \"472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9\": rpc error: code = NotFound desc = could not find container \"472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9\": container with ID starting with 472297e7cc2dd566396de007c9b6e5c875846fe9f4e17c7120fcd11877d732e9 not found: ID does not exist" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.331760 4756 scope.go:117] "RemoveContainer" containerID="9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f" Feb 03 09:38:33 crc kubenswrapper[4756]: E0203 09:38:33.338943 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f\": container with ID starting with 9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f not found: ID does not exist" containerID="9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.338984 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f"} err="failed to get container status \"9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f\": rpc error: code = NotFound desc = could not find container \"9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f\": container with ID starting with 9af8d335f8e8667faae14fa0fb2544a1c42f21eefc2949253b8ce952c315dd0f not found: ID does not exist" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.339006 4756 scope.go:117] "RemoveContainer" containerID="9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6" Feb 03 09:38:33 crc kubenswrapper[4756]: E0203 09:38:33.339317 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6\": container with ID starting with 9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6 not found: ID does not exist" containerID="9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.339349 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6"} err="failed to get container status \"9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6\": rpc error: code = NotFound desc = could not find container \"9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6\": container with ID starting with 9e426f80f791bd95923467d30e7a41e5a995b166bfaac00efb7900de144118c6 not found: ID does not exist" Feb 03 09:38:33 crc kubenswrapper[4756]: I0203 09:38:33.626665 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" path="/var/lib/kubelet/pods/31805a74-ec1e-4f9d-806d-5e77fe88f709/volumes" Feb 03 09:38:35 crc kubenswrapper[4756]: I0203 09:38:35.038720 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-7dmqr"] Feb 03 09:38:35 crc kubenswrapper[4756]: I0203 09:38:35.063127 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-7dmqr"] Feb 03 09:38:35 crc kubenswrapper[4756]: I0203 09:38:35.623396 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79b123a2-b4b7-4233-98e5-538664495935" path="/var/lib/kubelet/pods/79b123a2-b4b7-4233-98e5-538664495935/volumes" Feb 03 09:38:40 crc kubenswrapper[4756]: I0203 09:38:40.057513 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-x9wrg"] Feb 03 09:38:40 crc kubenswrapper[4756]: I0203 09:38:40.072026 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-x9wrg"] Feb 03 09:38:41 crc kubenswrapper[4756]: I0203 09:38:41.625116 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45100e47-7ef9-499c-b831-ea2994af9403" path="/var/lib/kubelet/pods/45100e47-7ef9-499c-b831-ea2994af9403/volumes" Feb 03 09:38:44 crc kubenswrapper[4756]: I0203 09:38:44.318326 4756 generic.go:334] "Generic (PLEG): container finished" podID="3c14e04f-9c29-45ae-84dc-bad543ff55d2" containerID="56133ba6daf6f471702b88ee8b5f84d58f1a6c7f09f8a21ff8dadda1fa13bf40" exitCode=0 Feb 03 09:38:44 crc kubenswrapper[4756]: I0203 09:38:44.318373 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" event={"ID":"3c14e04f-9c29-45ae-84dc-bad543ff55d2","Type":"ContainerDied","Data":"56133ba6daf6f471702b88ee8b5f84d58f1a6c7f09f8a21ff8dadda1fa13bf40"} Feb 03 09:38:44 crc kubenswrapper[4756]: I0203 09:38:44.614902 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:38:44 crc kubenswrapper[4756]: E0203 09:38:44.615258 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.736657 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.778541 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnmnv\" (UniqueName: \"kubernetes.io/projected/3c14e04f-9c29-45ae-84dc-bad543ff55d2-kube-api-access-dnmnv\") pod \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.778797 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-inventory\") pod \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.778996 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-ssh-key-openstack-edpm-ipam\") pod \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\" (UID: \"3c14e04f-9c29-45ae-84dc-bad543ff55d2\") " Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.784123 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c14e04f-9c29-45ae-84dc-bad543ff55d2-kube-api-access-dnmnv" (OuterVolumeSpecName: "kube-api-access-dnmnv") pod "3c14e04f-9c29-45ae-84dc-bad543ff55d2" (UID: "3c14e04f-9c29-45ae-84dc-bad543ff55d2"). InnerVolumeSpecName "kube-api-access-dnmnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.807782 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-inventory" (OuterVolumeSpecName: "inventory") pod "3c14e04f-9c29-45ae-84dc-bad543ff55d2" (UID: "3c14e04f-9c29-45ae-84dc-bad543ff55d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.810344 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3c14e04f-9c29-45ae-84dc-bad543ff55d2" (UID: "3c14e04f-9c29-45ae-84dc-bad543ff55d2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.881244 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnmnv\" (UniqueName: \"kubernetes.io/projected/3c14e04f-9c29-45ae-84dc-bad543ff55d2-kube-api-access-dnmnv\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.881282 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:45 crc kubenswrapper[4756]: I0203 09:38:45.881294 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c14e04f-9c29-45ae-84dc-bad543ff55d2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.337277 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" event={"ID":"3c14e04f-9c29-45ae-84dc-bad543ff55d2","Type":"ContainerDied","Data":"35bb1967f23b490db391aed68589c6942ff349a028becb23af2c2e31bf4011b1"} Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.337328 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35bb1967f23b490db391aed68589c6942ff349a028becb23af2c2e31bf4011b1" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.337486 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.421889 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc"] Feb 03 09:38:46 crc kubenswrapper[4756]: E0203 09:38:46.422234 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerName="registry-server" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.422247 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerName="registry-server" Feb 03 09:38:46 crc kubenswrapper[4756]: E0203 09:38:46.422271 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerName="extract-utilities" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.422278 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerName="extract-utilities" Feb 03 09:38:46 crc kubenswrapper[4756]: E0203 09:38:46.422298 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c14e04f-9c29-45ae-84dc-bad543ff55d2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.422306 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c14e04f-9c29-45ae-84dc-bad543ff55d2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 03 09:38:46 crc kubenswrapper[4756]: E0203 09:38:46.422320 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerName="extract-content" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.422327 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerName="extract-content" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.422557 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="31805a74-ec1e-4f9d-806d-5e77fe88f709" containerName="registry-server" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.422581 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c14e04f-9c29-45ae-84dc-bad543ff55d2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.423126 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.426301 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.426337 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.430701 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.436583 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.440321 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc"] Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.491520 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.491581 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9lcc\" (UniqueName: \"kubernetes.io/projected/4a58ee0e-41e0-4040-99c1-3261c3416a29-kube-api-access-w9lcc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.491703 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.592903 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.592962 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9lcc\" (UniqueName: \"kubernetes.io/projected/4a58ee0e-41e0-4040-99c1-3261c3416a29-kube-api-access-w9lcc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.593053 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.597158 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.597192 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.612100 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9lcc\" (UniqueName: \"kubernetes.io/projected/4a58ee0e-41e0-4040-99c1-3261c3416a29-kube-api-access-w9lcc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:46 crc kubenswrapper[4756]: I0203 09:38:46.748903 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:47 crc kubenswrapper[4756]: I0203 09:38:47.253128 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc"] Feb 03 09:38:47 crc kubenswrapper[4756]: I0203 09:38:47.354141 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" event={"ID":"4a58ee0e-41e0-4040-99c1-3261c3416a29","Type":"ContainerStarted","Data":"2428fc7d7e5f74e78258a8d6fc9837492da34f0e399dca9e7f66fae8788c98c3"} Feb 03 09:38:48 crc kubenswrapper[4756]: I0203 09:38:48.386389 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" event={"ID":"4a58ee0e-41e0-4040-99c1-3261c3416a29","Type":"ContainerStarted","Data":"6c88a32010b4e92e2c0f773ca003d9e02d04839cf466c01550126c755e0a0b15"} Feb 03 09:38:48 crc kubenswrapper[4756]: I0203 09:38:48.407230 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" podStartSLOduration=2.236032307 podStartE2EDuration="2.407213298s" podCreationTimestamp="2026-02-03 09:38:46 +0000 UTC" firstStartedPulling="2026-02-03 09:38:47.263806305 +0000 UTC m=+1718.414273680" lastFinishedPulling="2026-02-03 09:38:47.434987296 +0000 UTC m=+1718.585454671" observedRunningTime="2026-02-03 09:38:48.40279769 +0000 UTC m=+1719.553265105" watchObservedRunningTime="2026-02-03 09:38:48.407213298 +0000 UTC m=+1719.557680673" Feb 03 09:38:51 crc kubenswrapper[4756]: I0203 09:38:51.039929 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-4ld7v"] Feb 03 09:38:51 crc kubenswrapper[4756]: I0203 09:38:51.048917 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-cbcxp"] Feb 03 09:38:51 crc kubenswrapper[4756]: I0203 09:38:51.057704 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-4ld7v"] Feb 03 09:38:51 crc kubenswrapper[4756]: I0203 09:38:51.065406 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-cbcxp"] Feb 03 09:38:51 crc kubenswrapper[4756]: I0203 09:38:51.626129 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d6582a5-c7dc-4444-924e-9b977a3d4378" path="/var/lib/kubelet/pods/1d6582a5-c7dc-4444-924e-9b977a3d4378/volumes" Feb 03 09:38:51 crc kubenswrapper[4756]: I0203 09:38:51.626802 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dee8740-0310-450c-88fd-363ce0a26d4b" path="/var/lib/kubelet/pods/7dee8740-0310-450c-88fd-363ce0a26d4b/volumes" Feb 03 09:38:52 crc kubenswrapper[4756]: I0203 09:38:52.416810 4756 generic.go:334] "Generic (PLEG): container finished" podID="4a58ee0e-41e0-4040-99c1-3261c3416a29" containerID="6c88a32010b4e92e2c0f773ca003d9e02d04839cf466c01550126c755e0a0b15" exitCode=0 Feb 03 09:38:52 crc kubenswrapper[4756]: I0203 09:38:52.416906 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" event={"ID":"4a58ee0e-41e0-4040-99c1-3261c3416a29","Type":"ContainerDied","Data":"6c88a32010b4e92e2c0f773ca003d9e02d04839cf466c01550126c755e0a0b15"} Feb 03 09:38:53 crc kubenswrapper[4756]: I0203 09:38:53.848625 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:53 crc kubenswrapper[4756]: I0203 09:38:53.919128 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-inventory\") pod \"4a58ee0e-41e0-4040-99c1-3261c3416a29\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " Feb 03 09:38:53 crc kubenswrapper[4756]: I0203 09:38:53.919198 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9lcc\" (UniqueName: \"kubernetes.io/projected/4a58ee0e-41e0-4040-99c1-3261c3416a29-kube-api-access-w9lcc\") pod \"4a58ee0e-41e0-4040-99c1-3261c3416a29\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " Feb 03 09:38:53 crc kubenswrapper[4756]: I0203 09:38:53.919280 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-ssh-key-openstack-edpm-ipam\") pod \"4a58ee0e-41e0-4040-99c1-3261c3416a29\" (UID: \"4a58ee0e-41e0-4040-99c1-3261c3416a29\") " Feb 03 09:38:53 crc kubenswrapper[4756]: I0203 09:38:53.925696 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a58ee0e-41e0-4040-99c1-3261c3416a29-kube-api-access-w9lcc" (OuterVolumeSpecName: "kube-api-access-w9lcc") pod "4a58ee0e-41e0-4040-99c1-3261c3416a29" (UID: "4a58ee0e-41e0-4040-99c1-3261c3416a29"). InnerVolumeSpecName "kube-api-access-w9lcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:38:53 crc kubenswrapper[4756]: I0203 09:38:53.944460 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4a58ee0e-41e0-4040-99c1-3261c3416a29" (UID: "4a58ee0e-41e0-4040-99c1-3261c3416a29"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:38:53 crc kubenswrapper[4756]: I0203 09:38:53.946534 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-inventory" (OuterVolumeSpecName: "inventory") pod "4a58ee0e-41e0-4040-99c1-3261c3416a29" (UID: "4a58ee0e-41e0-4040-99c1-3261c3416a29"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.021627 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.021668 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9lcc\" (UniqueName: \"kubernetes.io/projected/4a58ee0e-41e0-4040-99c1-3261c3416a29-kube-api-access-w9lcc\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.021680 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4a58ee0e-41e0-4040-99c1-3261c3416a29-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.437039 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" event={"ID":"4a58ee0e-41e0-4040-99c1-3261c3416a29","Type":"ContainerDied","Data":"2428fc7d7e5f74e78258a8d6fc9837492da34f0e399dca9e7f66fae8788c98c3"} Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.437090 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2428fc7d7e5f74e78258a8d6fc9837492da34f0e399dca9e7f66fae8788c98c3" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.437115 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.519496 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9"] Feb 03 09:38:54 crc kubenswrapper[4756]: E0203 09:38:54.520007 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a58ee0e-41e0-4040-99c1-3261c3416a29" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.520029 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a58ee0e-41e0-4040-99c1-3261c3416a29" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.520245 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a58ee0e-41e0-4040-99c1-3261c3416a29" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.520965 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.522888 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.523210 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.523467 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.524260 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.528055 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9"] Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.630561 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.630647 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.630799 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sqmv\" (UniqueName: \"kubernetes.io/projected/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-kube-api-access-4sqmv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.732639 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sqmv\" (UniqueName: \"kubernetes.io/projected/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-kube-api-access-4sqmv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.733042 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.733081 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.739404 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.739988 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.756210 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sqmv\" (UniqueName: \"kubernetes.io/projected/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-kube-api-access-4sqmv\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-q5fk9\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:54 crc kubenswrapper[4756]: I0203 09:38:54.889099 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:38:55 crc kubenswrapper[4756]: I0203 09:38:55.409979 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9"] Feb 03 09:38:55 crc kubenswrapper[4756]: I0203 09:38:55.450801 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" event={"ID":"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad","Type":"ContainerStarted","Data":"b6d14f82539062e568db377281d51566d1f96355cbe057574fb65ab00ec1d3e7"} Feb 03 09:38:56 crc kubenswrapper[4756]: I0203 09:38:56.461679 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" event={"ID":"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad","Type":"ContainerStarted","Data":"217e5991b63d1ddacb64eb40b57766006c58b891aeb0165958208d05110338d1"} Feb 03 09:38:59 crc kubenswrapper[4756]: I0203 09:38:59.623369 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:38:59 crc kubenswrapper[4756]: E0203 09:38:59.624182 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:39:11 crc kubenswrapper[4756]: I0203 09:39:11.615056 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:39:11 crc kubenswrapper[4756]: E0203 09:39:11.615837 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:39:25 crc kubenswrapper[4756]: I0203 09:39:25.615230 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:39:25 crc kubenswrapper[4756]: E0203 09:39:25.616898 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:39:27 crc kubenswrapper[4756]: I0203 09:39:27.730414 4756 generic.go:334] "Generic (PLEG): container finished" podID="bf98eb35-81c9-4cc5-9d5c-d81d97d716ad" containerID="217e5991b63d1ddacb64eb40b57766006c58b891aeb0165958208d05110338d1" exitCode=0 Feb 03 09:39:27 crc kubenswrapper[4756]: I0203 09:39:27.730511 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" event={"ID":"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad","Type":"ContainerDied","Data":"217e5991b63d1ddacb64eb40b57766006c58b891aeb0165958208d05110338d1"} Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.174561 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.258298 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-inventory\") pod \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.258537 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sqmv\" (UniqueName: \"kubernetes.io/projected/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-kube-api-access-4sqmv\") pod \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.258557 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-ssh-key-openstack-edpm-ipam\") pod \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\" (UID: \"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad\") " Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.263924 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-kube-api-access-4sqmv" (OuterVolumeSpecName: "kube-api-access-4sqmv") pod "bf98eb35-81c9-4cc5-9d5c-d81d97d716ad" (UID: "bf98eb35-81c9-4cc5-9d5c-d81d97d716ad"). InnerVolumeSpecName "kube-api-access-4sqmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.283169 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bf98eb35-81c9-4cc5-9d5c-d81d97d716ad" (UID: "bf98eb35-81c9-4cc5-9d5c-d81d97d716ad"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.285962 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-inventory" (OuterVolumeSpecName: "inventory") pod "bf98eb35-81c9-4cc5-9d5c-d81d97d716ad" (UID: "bf98eb35-81c9-4cc5-9d5c-d81d97d716ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.361192 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sqmv\" (UniqueName: \"kubernetes.io/projected/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-kube-api-access-4sqmv\") on node \"crc\" DevicePath \"\"" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.361243 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.361263 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf98eb35-81c9-4cc5-9d5c-d81d97d716ad-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.753826 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" event={"ID":"bf98eb35-81c9-4cc5-9d5c-d81d97d716ad","Type":"ContainerDied","Data":"b6d14f82539062e568db377281d51566d1f96355cbe057574fb65ab00ec1d3e7"} Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.753864 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6d14f82539062e568db377281d51566d1f96355cbe057574fb65ab00ec1d3e7" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.753919 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-q5fk9" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.838143 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj"] Feb 03 09:39:29 crc kubenswrapper[4756]: E0203 09:39:29.839264 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf98eb35-81c9-4cc5-9d5c-d81d97d716ad" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.839374 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf98eb35-81c9-4cc5-9d5c-d81d97d716ad" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.839633 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf98eb35-81c9-4cc5-9d5c-d81d97d716ad" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.841828 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.854430 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj"] Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.883567 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.884259 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.884514 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.884690 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.990886 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtcjb\" (UniqueName: \"kubernetes.io/projected/9851b3a8-4288-4c0b-9ae9-38a5c043682f-kube-api-access-qtcjb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.990995 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:29 crc kubenswrapper[4756]: I0203 09:39:29.991029 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.042299 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-6s59j"] Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.050730 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-6s59j"] Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.093064 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.093105 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.093226 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtcjb\" (UniqueName: \"kubernetes.io/projected/9851b3a8-4288-4c0b-9ae9-38a5c043682f-kube-api-access-qtcjb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.098646 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.104796 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.122235 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtcjb\" (UniqueName: \"kubernetes.io/projected/9851b3a8-4288-4c0b-9ae9-38a5c043682f-kube-api-access-qtcjb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.195149 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.527386 4756 scope.go:117] "RemoveContainer" containerID="0439d59a66669486fb155370afc9837c626a409281f0397983d81cb77366332d" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.561973 4756 scope.go:117] "RemoveContainer" containerID="66e2f1708ffff8fa5abe63648080d850074bd13d775bd047ec3c99fca2a25235" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.620553 4756 scope.go:117] "RemoveContainer" containerID="5fe5f459c0dd48a21d9f0786435c92fa5dad2a769a3b44126c85575ac0fd1558" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.658762 4756 scope.go:117] "RemoveContainer" containerID="77986f06d0b2cfa7481468c9ca08776ae2dcce0ceb8ccdc36b9276a581777f7c" Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.703836 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj"] Feb 03 09:39:30 crc kubenswrapper[4756]: I0203 09:39:30.769402 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" event={"ID":"9851b3a8-4288-4c0b-9ae9-38a5c043682f","Type":"ContainerStarted","Data":"087c2c0a17e568f73ce9b6c9ff5178321eb4aebbfff7f5e231ac27f1088ec54c"} Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.031226 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-d2m25"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.041719 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-d2m25"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.051166 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5fd1-account-create-update-6kqzn"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.061657 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-x676l"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.070144 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f912-account-create-update-t2sfb"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.078026 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5fd1-account-create-update-6kqzn"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.088097 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f912-account-create-update-t2sfb"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.095391 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f9ed-account-create-update-jnkxm"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.103178 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-x676l"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.111034 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f9ed-account-create-update-jnkxm"] Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.624644 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15a12c2e-5874-45c7-bc27-1281125fb024" path="/var/lib/kubelet/pods/15a12c2e-5874-45c7-bc27-1281125fb024/volumes" Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.625616 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24066fdc-6998-4f09-9a07-230cb3378f11" path="/var/lib/kubelet/pods/24066fdc-6998-4f09-9a07-230cb3378f11/volumes" Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.626100 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4172aecf-8ba4-4990-9bd9-11ef49b0320e" path="/var/lib/kubelet/pods/4172aecf-8ba4-4990-9bd9-11ef49b0320e/volumes" Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.627014 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95eac2ad-73dc-405d-b583-92d263bec833" path="/var/lib/kubelet/pods/95eac2ad-73dc-405d-b583-92d263bec833/volumes" Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.629208 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b922d8-73c3-42e7-a7af-f28120f67125" path="/var/lib/kubelet/pods/a2b922d8-73c3-42e7-a7af-f28120f67125/volumes" Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.629737 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea" path="/var/lib/kubelet/pods/d2e9b443-1400-4bf7-b4c9-4cb3eba4f0ea/volumes" Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.779926 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" event={"ID":"9851b3a8-4288-4c0b-9ae9-38a5c043682f","Type":"ContainerStarted","Data":"db0ea58a570489677cfc99f468c2908a368651a77a59849320f7d3f2e9ddba00"} Feb 03 09:39:31 crc kubenswrapper[4756]: I0203 09:39:31.796910 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" podStartSLOduration=2.651812165 podStartE2EDuration="2.796890602s" podCreationTimestamp="2026-02-03 09:39:29 +0000 UTC" firstStartedPulling="2026-02-03 09:39:30.710732115 +0000 UTC m=+1761.861199490" lastFinishedPulling="2026-02-03 09:39:30.855810562 +0000 UTC m=+1762.006277927" observedRunningTime="2026-02-03 09:39:31.795041425 +0000 UTC m=+1762.945508800" watchObservedRunningTime="2026-02-03 09:39:31.796890602 +0000 UTC m=+1762.947357977" Feb 03 09:39:37 crc kubenswrapper[4756]: I0203 09:39:37.614842 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:39:37 crc kubenswrapper[4756]: E0203 09:39:37.615645 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:39:49 crc kubenswrapper[4756]: I0203 09:39:49.623116 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:39:49 crc kubenswrapper[4756]: E0203 09:39:49.624513 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:40:00 crc kubenswrapper[4756]: I0203 09:40:00.613826 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:40:00 crc kubenswrapper[4756]: E0203 09:40:00.614643 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:40:01 crc kubenswrapper[4756]: I0203 09:40:01.044310 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zkffm"] Feb 03 09:40:01 crc kubenswrapper[4756]: I0203 09:40:01.051997 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zkffm"] Feb 03 09:40:01 crc kubenswrapper[4756]: I0203 09:40:01.640601 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a34d89c-812e-4f86-8ec1-6a38cfd710c8" path="/var/lib/kubelet/pods/5a34d89c-812e-4f86-8ec1-6a38cfd710c8/volumes" Feb 03 09:40:11 crc kubenswrapper[4756]: I0203 09:40:11.614526 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:40:11 crc kubenswrapper[4756]: E0203 09:40:11.615894 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:40:15 crc kubenswrapper[4756]: I0203 09:40:15.146371 4756 generic.go:334] "Generic (PLEG): container finished" podID="9851b3a8-4288-4c0b-9ae9-38a5c043682f" containerID="db0ea58a570489677cfc99f468c2908a368651a77a59849320f7d3f2e9ddba00" exitCode=0 Feb 03 09:40:15 crc kubenswrapper[4756]: I0203 09:40:15.146476 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" event={"ID":"9851b3a8-4288-4c0b-9ae9-38a5c043682f","Type":"ContainerDied","Data":"db0ea58a570489677cfc99f468c2908a368651a77a59849320f7d3f2e9ddba00"} Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.563243 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.687259 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtcjb\" (UniqueName: \"kubernetes.io/projected/9851b3a8-4288-4c0b-9ae9-38a5c043682f-kube-api-access-qtcjb\") pod \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.687406 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-inventory\") pod \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.687694 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-ssh-key-openstack-edpm-ipam\") pod \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\" (UID: \"9851b3a8-4288-4c0b-9ae9-38a5c043682f\") " Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.692780 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9851b3a8-4288-4c0b-9ae9-38a5c043682f-kube-api-access-qtcjb" (OuterVolumeSpecName: "kube-api-access-qtcjb") pod "9851b3a8-4288-4c0b-9ae9-38a5c043682f" (UID: "9851b3a8-4288-4c0b-9ae9-38a5c043682f"). InnerVolumeSpecName "kube-api-access-qtcjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.713234 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9851b3a8-4288-4c0b-9ae9-38a5c043682f" (UID: "9851b3a8-4288-4c0b-9ae9-38a5c043682f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.713786 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-inventory" (OuterVolumeSpecName: "inventory") pod "9851b3a8-4288-4c0b-9ae9-38a5c043682f" (UID: "9851b3a8-4288-4c0b-9ae9-38a5c043682f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.790053 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.790084 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9851b3a8-4288-4c0b-9ae9-38a5c043682f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:16 crc kubenswrapper[4756]: I0203 09:40:16.790095 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtcjb\" (UniqueName: \"kubernetes.io/projected/9851b3a8-4288-4c0b-9ae9-38a5c043682f-kube-api-access-qtcjb\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.161904 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" event={"ID":"9851b3a8-4288-4c0b-9ae9-38a5c043682f","Type":"ContainerDied","Data":"087c2c0a17e568f73ce9b6c9ff5178321eb4aebbfff7f5e231ac27f1088ec54c"} Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.161943 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="087c2c0a17e568f73ce9b6c9ff5178321eb4aebbfff7f5e231ac27f1088ec54c" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.161948 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.242841 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7jsx4"] Feb 03 09:40:17 crc kubenswrapper[4756]: E0203 09:40:17.243247 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9851b3a8-4288-4c0b-9ae9-38a5c043682f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.243265 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="9851b3a8-4288-4c0b-9ae9-38a5c043682f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.243509 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="9851b3a8-4288-4c0b-9ae9-38a5c043682f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.244289 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.247017 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.247318 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.247593 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.247679 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.254080 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7jsx4"] Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.297359 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.297512 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.297592 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwr59\" (UniqueName: \"kubernetes.io/projected/848ce5a7-c89c-4cac-8275-263ca14eac78-kube-api-access-vwr59\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.399701 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwr59\" (UniqueName: \"kubernetes.io/projected/848ce5a7-c89c-4cac-8275-263ca14eac78-kube-api-access-vwr59\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.399805 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.399906 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.403754 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.404237 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.416814 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwr59\" (UniqueName: \"kubernetes.io/projected/848ce5a7-c89c-4cac-8275-263ca14eac78-kube-api-access-vwr59\") pod \"ssh-known-hosts-edpm-deployment-7jsx4\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:17 crc kubenswrapper[4756]: I0203 09:40:17.567469 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:18 crc kubenswrapper[4756]: I0203 09:40:18.038602 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-7kx6w"] Feb 03 09:40:18 crc kubenswrapper[4756]: I0203 09:40:18.050071 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-7kx6w"] Feb 03 09:40:18 crc kubenswrapper[4756]: I0203 09:40:18.072829 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-7jsx4"] Feb 03 09:40:18 crc kubenswrapper[4756]: I0203 09:40:18.170722 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" event={"ID":"848ce5a7-c89c-4cac-8275-263ca14eac78","Type":"ContainerStarted","Data":"b93c9dbec5750aba4674b4f3a255a510d4572b568a2b7479b66d4d971c29179b"} Feb 03 09:40:19 crc kubenswrapper[4756]: I0203 09:40:19.034086 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4qgbq"] Feb 03 09:40:19 crc kubenswrapper[4756]: I0203 09:40:19.045403 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-4qgbq"] Feb 03 09:40:19 crc kubenswrapper[4756]: I0203 09:40:19.183845 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" event={"ID":"848ce5a7-c89c-4cac-8275-263ca14eac78","Type":"ContainerStarted","Data":"1abaf2b89d59c54219b7b091a86a8152059708599b9a6a2520d5318e9fdfad6d"} Feb 03 09:40:19 crc kubenswrapper[4756]: I0203 09:40:19.215916 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" podStartSLOduration=2.012482138 podStartE2EDuration="2.215895723s" podCreationTimestamp="2026-02-03 09:40:17 +0000 UTC" firstStartedPulling="2026-02-03 09:40:18.074356979 +0000 UTC m=+1809.224824364" lastFinishedPulling="2026-02-03 09:40:18.277770574 +0000 UTC m=+1809.428237949" observedRunningTime="2026-02-03 09:40:19.208277055 +0000 UTC m=+1810.358744430" watchObservedRunningTime="2026-02-03 09:40:19.215895723 +0000 UTC m=+1810.366363098" Feb 03 09:40:19 crc kubenswrapper[4756]: I0203 09:40:19.624085 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="197062ac-2a3d-4b9b-a099-0d5843e4c58b" path="/var/lib/kubelet/pods/197062ac-2a3d-4b9b-a099-0d5843e4c58b/volumes" Feb 03 09:40:19 crc kubenswrapper[4756]: I0203 09:40:19.624666 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af01732-2a75-4a7d-af19-b097bc0ffd6b" path="/var/lib/kubelet/pods/5af01732-2a75-4a7d-af19-b097bc0ffd6b/volumes" Feb 03 09:40:25 crc kubenswrapper[4756]: I0203 09:40:25.243341 4756 generic.go:334] "Generic (PLEG): container finished" podID="848ce5a7-c89c-4cac-8275-263ca14eac78" containerID="1abaf2b89d59c54219b7b091a86a8152059708599b9a6a2520d5318e9fdfad6d" exitCode=0 Feb 03 09:40:25 crc kubenswrapper[4756]: I0203 09:40:25.243437 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" event={"ID":"848ce5a7-c89c-4cac-8275-263ca14eac78","Type":"ContainerDied","Data":"1abaf2b89d59c54219b7b091a86a8152059708599b9a6a2520d5318e9fdfad6d"} Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.613749 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.652534 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.687460 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-ssh-key-openstack-edpm-ipam\") pod \"848ce5a7-c89c-4cac-8275-263ca14eac78\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.687922 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-inventory-0\") pod \"848ce5a7-c89c-4cac-8275-263ca14eac78\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.687969 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwr59\" (UniqueName: \"kubernetes.io/projected/848ce5a7-c89c-4cac-8275-263ca14eac78-kube-api-access-vwr59\") pod \"848ce5a7-c89c-4cac-8275-263ca14eac78\" (UID: \"848ce5a7-c89c-4cac-8275-263ca14eac78\") " Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.694421 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/848ce5a7-c89c-4cac-8275-263ca14eac78-kube-api-access-vwr59" (OuterVolumeSpecName: "kube-api-access-vwr59") pod "848ce5a7-c89c-4cac-8275-263ca14eac78" (UID: "848ce5a7-c89c-4cac-8275-263ca14eac78"). InnerVolumeSpecName "kube-api-access-vwr59". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.718478 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "848ce5a7-c89c-4cac-8275-263ca14eac78" (UID: "848ce5a7-c89c-4cac-8275-263ca14eac78"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.724601 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "848ce5a7-c89c-4cac-8275-263ca14eac78" (UID: "848ce5a7-c89c-4cac-8275-263ca14eac78"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.790436 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.790482 4756 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/848ce5a7-c89c-4cac-8275-263ca14eac78-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:26 crc kubenswrapper[4756]: I0203 09:40:26.790492 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwr59\" (UniqueName: \"kubernetes.io/projected/848ce5a7-c89c-4cac-8275-263ca14eac78-kube-api-access-vwr59\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.259575 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" event={"ID":"848ce5a7-c89c-4cac-8275-263ca14eac78","Type":"ContainerDied","Data":"b93c9dbec5750aba4674b4f3a255a510d4572b568a2b7479b66d4d971c29179b"} Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.259624 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b93c9dbec5750aba4674b4f3a255a510d4572b568a2b7479b66d4d971c29179b" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.259664 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-7jsx4" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.261723 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"6e6c9e3d9602222af7b0316088c411a3230434432a9c9562494d24bc001e56b1"} Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.363555 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk"] Feb 03 09:40:27 crc kubenswrapper[4756]: E0203 09:40:27.364272 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="848ce5a7-c89c-4cac-8275-263ca14eac78" containerName="ssh-known-hosts-edpm-deployment" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.364289 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="848ce5a7-c89c-4cac-8275-263ca14eac78" containerName="ssh-known-hosts-edpm-deployment" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.364505 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="848ce5a7-c89c-4cac-8275-263ca14eac78" containerName="ssh-known-hosts-edpm-deployment" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.365796 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.368824 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.368829 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.373104 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk"] Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.374555 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.374583 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.400048 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.400119 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g2pq\" (UniqueName: \"kubernetes.io/projected/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-kube-api-access-2g2pq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.400265 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: E0203 09:40:27.402569 4756 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod848ce5a7_c89c_4cac_8275_263ca14eac78.slice/crio-b93c9dbec5750aba4674b4f3a255a510d4572b568a2b7479b66d4d971c29179b\": RecentStats: unable to find data in memory cache]" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.501751 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.501809 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g2pq\" (UniqueName: \"kubernetes.io/projected/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-kube-api-access-2g2pq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.501920 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.506746 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.507265 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.523930 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g2pq\" (UniqueName: \"kubernetes.io/projected/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-kube-api-access-2g2pq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9g7tk\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:27 crc kubenswrapper[4756]: I0203 09:40:27.686274 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:28 crc kubenswrapper[4756]: I0203 09:40:28.273452 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk"] Feb 03 09:40:28 crc kubenswrapper[4756]: W0203 09:40:28.276946 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5181cb59_0c71_4a78_8ba8_d5c5a22c4dbf.slice/crio-5cbed364145a599ea58489b08e405061f602ee0b208bff66289665b9e06615dd WatchSource:0}: Error finding container 5cbed364145a599ea58489b08e405061f602ee0b208bff66289665b9e06615dd: Status 404 returned error can't find the container with id 5cbed364145a599ea58489b08e405061f602ee0b208bff66289665b9e06615dd Feb 03 09:40:29 crc kubenswrapper[4756]: I0203 09:40:29.283161 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" event={"ID":"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf","Type":"ContainerStarted","Data":"2acbbc5efee7a1ea00f380ad2e6072e2b6390a768676fb3cac1c369a73468904"} Feb 03 09:40:29 crc kubenswrapper[4756]: I0203 09:40:29.283772 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" event={"ID":"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf","Type":"ContainerStarted","Data":"5cbed364145a599ea58489b08e405061f602ee0b208bff66289665b9e06615dd"} Feb 03 09:40:29 crc kubenswrapper[4756]: I0203 09:40:29.298383 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" podStartSLOduration=2.145157871 podStartE2EDuration="2.298362211s" podCreationTimestamp="2026-02-03 09:40:27 +0000 UTC" firstStartedPulling="2026-02-03 09:40:28.280073731 +0000 UTC m=+1819.430541106" lastFinishedPulling="2026-02-03 09:40:28.433278071 +0000 UTC m=+1819.583745446" observedRunningTime="2026-02-03 09:40:29.296929875 +0000 UTC m=+1820.447397250" watchObservedRunningTime="2026-02-03 09:40:29.298362211 +0000 UTC m=+1820.448829586" Feb 03 09:40:30 crc kubenswrapper[4756]: I0203 09:40:30.793921 4756 scope.go:117] "RemoveContainer" containerID="65183e148b041543009d86cf8844f0a62e8424d31830d5c112de1d6df413fccd" Feb 03 09:40:30 crc kubenswrapper[4756]: I0203 09:40:30.829333 4756 scope.go:117] "RemoveContainer" containerID="67abb070992b5549ffeba6a615e778eb71f9e9e5e9cf9624ab62546df59517b2" Feb 03 09:40:30 crc kubenswrapper[4756]: I0203 09:40:30.862616 4756 scope.go:117] "RemoveContainer" containerID="aad8c853f8c178c99a92838cbe7181f9283f11434ba6a9a3593f17c77d43dabe" Feb 03 09:40:30 crc kubenswrapper[4756]: I0203 09:40:30.911749 4756 scope.go:117] "RemoveContainer" containerID="23531e769d889687f55e0157730d72e4aaae9107e14ac8e940390a190adeafbe" Feb 03 09:40:30 crc kubenswrapper[4756]: I0203 09:40:30.965800 4756 scope.go:117] "RemoveContainer" containerID="d2f60bbc9e7f4bbd60b771f014093defccf5e361ff795b00ee4e450d490222d0" Feb 03 09:40:30 crc kubenswrapper[4756]: I0203 09:40:30.990632 4756 scope.go:117] "RemoveContainer" containerID="a654d32d14744b248ff72832f42e1ca11792dd1ab1ab25b9372dcfafe604a53d" Feb 03 09:40:31 crc kubenswrapper[4756]: I0203 09:40:31.038480 4756 scope.go:117] "RemoveContainer" containerID="81c525f541f93ae9180d083bfaf1b9f216fb3827bccee3cccc00d04cebf6a0ca" Feb 03 09:40:31 crc kubenswrapper[4756]: I0203 09:40:31.088906 4756 scope.go:117] "RemoveContainer" containerID="2fdbf7fc7b72893e059e77d00d4c50b542d97835f099d9f579047043b519cb79" Feb 03 09:40:31 crc kubenswrapper[4756]: I0203 09:40:31.107152 4756 scope.go:117] "RemoveContainer" containerID="35ceb7ea9786f3b105817c46c6f20f572d4769869bb13a1ce23f96fc639c759c" Feb 03 09:40:36 crc kubenswrapper[4756]: I0203 09:40:36.339274 4756 generic.go:334] "Generic (PLEG): container finished" podID="5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf" containerID="2acbbc5efee7a1ea00f380ad2e6072e2b6390a768676fb3cac1c369a73468904" exitCode=0 Feb 03 09:40:36 crc kubenswrapper[4756]: I0203 09:40:36.339368 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" event={"ID":"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf","Type":"ContainerDied","Data":"2acbbc5efee7a1ea00f380ad2e6072e2b6390a768676fb3cac1c369a73468904"} Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.763350 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.822949 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g2pq\" (UniqueName: \"kubernetes.io/projected/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-kube-api-access-2g2pq\") pod \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.823002 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-ssh-key-openstack-edpm-ipam\") pod \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.823177 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-inventory\") pod \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\" (UID: \"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf\") " Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.828580 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-kube-api-access-2g2pq" (OuterVolumeSpecName: "kube-api-access-2g2pq") pod "5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf" (UID: "5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf"). InnerVolumeSpecName "kube-api-access-2g2pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.849991 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf" (UID: "5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.852027 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-inventory" (OuterVolumeSpecName: "inventory") pod "5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf" (UID: "5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.925614 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.925651 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g2pq\" (UniqueName: \"kubernetes.io/projected/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-kube-api-access-2g2pq\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:37 crc kubenswrapper[4756]: I0203 09:40:37.925663 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.359105 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" event={"ID":"5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf","Type":"ContainerDied","Data":"5cbed364145a599ea58489b08e405061f602ee0b208bff66289665b9e06615dd"} Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.359482 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cbed364145a599ea58489b08e405061f602ee0b208bff66289665b9e06615dd" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.359145 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9g7tk" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.420123 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw"] Feb 03 09:40:38 crc kubenswrapper[4756]: E0203 09:40:38.420530 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.420550 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.420716 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.421325 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.424512 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.425345 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.425545 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.425715 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.440588 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw"] Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.535391 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.535511 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vclt\" (UniqueName: \"kubernetes.io/projected/4ae5e9e1-d148-4487-862a-b2312222f2b0-kube-api-access-4vclt\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.535987 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.638162 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.638237 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.638271 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vclt\" (UniqueName: \"kubernetes.io/projected/4ae5e9e1-d148-4487-862a-b2312222f2b0-kube-api-access-4vclt\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.642848 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.643608 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.657649 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vclt\" (UniqueName: \"kubernetes.io/projected/4ae5e9e1-d148-4487-862a-b2312222f2b0-kube-api-access-4vclt\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:38 crc kubenswrapper[4756]: I0203 09:40:38.744578 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:39 crc kubenswrapper[4756]: I0203 09:40:39.306977 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw"] Feb 03 09:40:39 crc kubenswrapper[4756]: I0203 09:40:39.366489 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" event={"ID":"4ae5e9e1-d148-4487-862a-b2312222f2b0","Type":"ContainerStarted","Data":"df040a5b02c131c277f490b39e2b186427e9f17c05fbc6731c19bbbd91c0f6e9"} Feb 03 09:40:40 crc kubenswrapper[4756]: I0203 09:40:40.377329 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" event={"ID":"4ae5e9e1-d148-4487-862a-b2312222f2b0","Type":"ContainerStarted","Data":"f64d722fa7ef8778eac8a2ecb12533d96275d21ae5a8668fafb1653e05b649fa"} Feb 03 09:40:40 crc kubenswrapper[4756]: I0203 09:40:40.402042 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" podStartSLOduration=2.201499283 podStartE2EDuration="2.402023689s" podCreationTimestamp="2026-02-03 09:40:38 +0000 UTC" firstStartedPulling="2026-02-03 09:40:39.322651573 +0000 UTC m=+1830.473118948" lastFinishedPulling="2026-02-03 09:40:39.523175979 +0000 UTC m=+1830.673643354" observedRunningTime="2026-02-03 09:40:40.396856437 +0000 UTC m=+1831.547323822" watchObservedRunningTime="2026-02-03 09:40:40.402023689 +0000 UTC m=+1831.552491064" Feb 03 09:40:48 crc kubenswrapper[4756]: I0203 09:40:48.442002 4756 generic.go:334] "Generic (PLEG): container finished" podID="4ae5e9e1-d148-4487-862a-b2312222f2b0" containerID="f64d722fa7ef8778eac8a2ecb12533d96275d21ae5a8668fafb1653e05b649fa" exitCode=0 Feb 03 09:40:48 crc kubenswrapper[4756]: I0203 09:40:48.442077 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" event={"ID":"4ae5e9e1-d148-4487-862a-b2312222f2b0","Type":"ContainerDied","Data":"f64d722fa7ef8778eac8a2ecb12533d96275d21ae5a8668fafb1653e05b649fa"} Feb 03 09:40:49 crc kubenswrapper[4756]: I0203 09:40:49.876650 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:49 crc kubenswrapper[4756]: I0203 09:40:49.971044 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-ssh-key-openstack-edpm-ipam\") pod \"4ae5e9e1-d148-4487-862a-b2312222f2b0\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " Feb 03 09:40:49 crc kubenswrapper[4756]: I0203 09:40:49.971179 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-inventory\") pod \"4ae5e9e1-d148-4487-862a-b2312222f2b0\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " Feb 03 09:40:49 crc kubenswrapper[4756]: I0203 09:40:49.971325 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vclt\" (UniqueName: \"kubernetes.io/projected/4ae5e9e1-d148-4487-862a-b2312222f2b0-kube-api-access-4vclt\") pod \"4ae5e9e1-d148-4487-862a-b2312222f2b0\" (UID: \"4ae5e9e1-d148-4487-862a-b2312222f2b0\") " Feb 03 09:40:49 crc kubenswrapper[4756]: I0203 09:40:49.977667 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ae5e9e1-d148-4487-862a-b2312222f2b0-kube-api-access-4vclt" (OuterVolumeSpecName: "kube-api-access-4vclt") pod "4ae5e9e1-d148-4487-862a-b2312222f2b0" (UID: "4ae5e9e1-d148-4487-862a-b2312222f2b0"). InnerVolumeSpecName "kube-api-access-4vclt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:40:49 crc kubenswrapper[4756]: I0203 09:40:49.997783 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-inventory" (OuterVolumeSpecName: "inventory") pod "4ae5e9e1-d148-4487-862a-b2312222f2b0" (UID: "4ae5e9e1-d148-4487-862a-b2312222f2b0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.014067 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4ae5e9e1-d148-4487-862a-b2312222f2b0" (UID: "4ae5e9e1-d148-4487-862a-b2312222f2b0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.074885 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.074950 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vclt\" (UniqueName: \"kubernetes.io/projected/4ae5e9e1-d148-4487-862a-b2312222f2b0-kube-api-access-4vclt\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.074964 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4ae5e9e1-d148-4487-862a-b2312222f2b0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.458005 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" event={"ID":"4ae5e9e1-d148-4487-862a-b2312222f2b0","Type":"ContainerDied","Data":"df040a5b02c131c277f490b39e2b186427e9f17c05fbc6731c19bbbd91c0f6e9"} Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.458043 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df040a5b02c131c277f490b39e2b186427e9f17c05fbc6731c19bbbd91c0f6e9" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.458062 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.547986 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw"] Feb 03 09:40:50 crc kubenswrapper[4756]: E0203 09:40:50.548425 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae5e9e1-d148-4487-862a-b2312222f2b0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.548467 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae5e9e1-d148-4487-862a-b2312222f2b0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.548722 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ae5e9e1-d148-4487-862a-b2312222f2b0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.549472 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.556923 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.557106 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.557291 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.557461 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.557610 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.557753 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.557907 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.564315 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw"] Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.574563 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.685225 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.685278 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.685435 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.685520 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.685708 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.685799 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.686016 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.686059 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.686121 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.686297 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.686364 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.686411 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.686453 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.686592 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9slwr\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-kube-api-access-9slwr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.788566 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.788672 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.788820 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9slwr\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-kube-api-access-9slwr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.789434 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790259 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790422 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790518 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790601 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790665 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790721 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790781 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790844 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.790937 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.791033 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.793925 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.794193 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.794258 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.794340 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.795021 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.795267 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.795973 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.797020 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.797059 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.797381 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.797404 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.797635 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.798186 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.807398 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9slwr\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-kube-api-access-9slwr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-9nclw\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:50 crc kubenswrapper[4756]: I0203 09:40:50.914919 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:40:51 crc kubenswrapper[4756]: I0203 09:40:51.549022 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw"] Feb 03 09:40:51 crc kubenswrapper[4756]: W0203 09:40:51.551193 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2396060c_bdc9_4f9c_8a42_c86984922418.slice/crio-105a56746f97cd1d7f501b24493e50644a0ff1cda3154c18d088559893076747 WatchSource:0}: Error finding container 105a56746f97cd1d7f501b24493e50644a0ff1cda3154c18d088559893076747: Status 404 returned error can't find the container with id 105a56746f97cd1d7f501b24493e50644a0ff1cda3154c18d088559893076747 Feb 03 09:40:52 crc kubenswrapper[4756]: I0203 09:40:52.475432 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" event={"ID":"2396060c-bdc9-4f9c-8a42-c86984922418","Type":"ContainerStarted","Data":"f688c598634d0055880c732ead3ee57f5d75b9adef5769adceb7a98f5128bcf9"} Feb 03 09:40:52 crc kubenswrapper[4756]: I0203 09:40:52.475753 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" event={"ID":"2396060c-bdc9-4f9c-8a42-c86984922418","Type":"ContainerStarted","Data":"105a56746f97cd1d7f501b24493e50644a0ff1cda3154c18d088559893076747"} Feb 03 09:40:52 crc kubenswrapper[4756]: I0203 09:40:52.494665 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" podStartSLOduration=2.324385489 podStartE2EDuration="2.4946481s" podCreationTimestamp="2026-02-03 09:40:50 +0000 UTC" firstStartedPulling="2026-02-03 09:40:51.554078626 +0000 UTC m=+1842.704546001" lastFinishedPulling="2026-02-03 09:40:51.724341237 +0000 UTC m=+1842.874808612" observedRunningTime="2026-02-03 09:40:52.49177229 +0000 UTC m=+1843.642239665" watchObservedRunningTime="2026-02-03 09:40:52.4946481 +0000 UTC m=+1843.645115475" Feb 03 09:41:04 crc kubenswrapper[4756]: I0203 09:41:04.041007 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mw4b6"] Feb 03 09:41:04 crc kubenswrapper[4756]: I0203 09:41:04.051115 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mw4b6"] Feb 03 09:41:06 crc kubenswrapper[4756]: I0203 09:41:06.859094 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50d3d3a0-68e5-449d-94fb-b295fd0d073e" path="/var/lib/kubelet/pods/50d3d3a0-68e5-449d-94fb-b295fd0d073e/volumes" Feb 03 09:41:26 crc kubenswrapper[4756]: I0203 09:41:26.000913 4756 generic.go:334] "Generic (PLEG): container finished" podID="2396060c-bdc9-4f9c-8a42-c86984922418" containerID="f688c598634d0055880c732ead3ee57f5d75b9adef5769adceb7a98f5128bcf9" exitCode=0 Feb 03 09:41:26 crc kubenswrapper[4756]: I0203 09:41:26.001664 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" event={"ID":"2396060c-bdc9-4f9c-8a42-c86984922418","Type":"ContainerDied","Data":"f688c598634d0055880c732ead3ee57f5d75b9adef5769adceb7a98f5128bcf9"} Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.469514 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634244 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634284 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634344 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-neutron-metadata-combined-ca-bundle\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634369 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-nova-combined-ca-bundle\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634481 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634508 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-libvirt-combined-ca-bundle\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634528 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9slwr\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-kube-api-access-9slwr\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634608 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-ovn-default-certs-0\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634648 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-bootstrap-combined-ca-bundle\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634680 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-telemetry-combined-ca-bundle\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634747 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ssh-key-openstack-edpm-ipam\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634788 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ovn-combined-ca-bundle\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634818 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-repo-setup-combined-ca-bundle\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.634870 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-inventory\") pod \"2396060c-bdc9-4f9c-8a42-c86984922418\" (UID: \"2396060c-bdc9-4f9c-8a42-c86984922418\") " Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.642637 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.642872 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.643438 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.643480 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.643836 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.643953 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.644077 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.644051 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.644705 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.647186 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-kube-api-access-9slwr" (OuterVolumeSpecName: "kube-api-access-9slwr") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "kube-api-access-9slwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.647402 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.650640 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.667765 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.681728 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-inventory" (OuterVolumeSpecName: "inventory") pod "2396060c-bdc9-4f9c-8a42-c86984922418" (UID: "2396060c-bdc9-4f9c-8a42-c86984922418"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737222 4756 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737263 4756 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737275 4756 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737286 4756 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737297 4756 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737306 4756 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737315 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9slwr\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-kube-api-access-9slwr\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737324 4756 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2396060c-bdc9-4f9c-8a42-c86984922418-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737334 4756 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737342 4756 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737351 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737360 4756 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737369 4756 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:27 crc kubenswrapper[4756]: I0203 09:41:27.737379 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2396060c-bdc9-4f9c-8a42-c86984922418-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.032677 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" event={"ID":"2396060c-bdc9-4f9c-8a42-c86984922418","Type":"ContainerDied","Data":"105a56746f97cd1d7f501b24493e50644a0ff1cda3154c18d088559893076747"} Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.032722 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="105a56746f97cd1d7f501b24493e50644a0ff1cda3154c18d088559893076747" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.032744 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-9nclw" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.130695 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l"] Feb 03 09:41:28 crc kubenswrapper[4756]: E0203 09:41:28.131166 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2396060c-bdc9-4f9c-8a42-c86984922418" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.131186 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="2396060c-bdc9-4f9c-8a42-c86984922418" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.131438 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="2396060c-bdc9-4f9c-8a42-c86984922418" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.132263 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.134315 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.134704 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.135306 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.135716 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.135907 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.141002 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l"] Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.249313 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c6bade3-63d5-4342-8ded-feaf11b54263-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.249537 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.249789 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p6mh\" (UniqueName: \"kubernetes.io/projected/3c6bade3-63d5-4342-8ded-feaf11b54263-kube-api-access-6p6mh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.249922 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.250060 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.351867 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.351943 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p6mh\" (UniqueName: \"kubernetes.io/projected/3c6bade3-63d5-4342-8ded-feaf11b54263-kube-api-access-6p6mh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.351982 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.352030 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.352087 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c6bade3-63d5-4342-8ded-feaf11b54263-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.352979 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c6bade3-63d5-4342-8ded-feaf11b54263-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.357109 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.357342 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.368807 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.371974 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p6mh\" (UniqueName: \"kubernetes.io/projected/3c6bade3-63d5-4342-8ded-feaf11b54263-kube-api-access-6p6mh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jbd5l\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.454787 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:41:28 crc kubenswrapper[4756]: I0203 09:41:28.955235 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l"] Feb 03 09:41:29 crc kubenswrapper[4756]: I0203 09:41:29.051435 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" event={"ID":"3c6bade3-63d5-4342-8ded-feaf11b54263","Type":"ContainerStarted","Data":"670374b188591e6166f19f8e3873145624ac3609ca4efc2a558846e046dfdb9d"} Feb 03 09:41:30 crc kubenswrapper[4756]: I0203 09:41:30.064028 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" event={"ID":"3c6bade3-63d5-4342-8ded-feaf11b54263","Type":"ContainerStarted","Data":"ecec43a9aa0edfd74c1bac14b8d026c06224210059018b414741ade266d698ca"} Feb 03 09:41:30 crc kubenswrapper[4756]: I0203 09:41:30.089847 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" podStartSLOduration=1.896028223 podStartE2EDuration="2.089821636s" podCreationTimestamp="2026-02-03 09:41:28 +0000 UTC" firstStartedPulling="2026-02-03 09:41:28.966219453 +0000 UTC m=+1880.116686828" lastFinishedPulling="2026-02-03 09:41:29.160012866 +0000 UTC m=+1880.310480241" observedRunningTime="2026-02-03 09:41:30.083261952 +0000 UTC m=+1881.233729327" watchObservedRunningTime="2026-02-03 09:41:30.089821636 +0000 UTC m=+1881.240289041" Feb 03 09:41:31 crc kubenswrapper[4756]: I0203 09:41:31.271099 4756 scope.go:117] "RemoveContainer" containerID="593f53ccf44138863664fff0bd20206942d11fe64a3a5e7f228ce19684693c23" Feb 03 09:42:26 crc kubenswrapper[4756]: I0203 09:42:26.613890 4756 generic.go:334] "Generic (PLEG): container finished" podID="3c6bade3-63d5-4342-8ded-feaf11b54263" containerID="ecec43a9aa0edfd74c1bac14b8d026c06224210059018b414741ade266d698ca" exitCode=0 Feb 03 09:42:26 crc kubenswrapper[4756]: I0203 09:42:26.613964 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" event={"ID":"3c6bade3-63d5-4342-8ded-feaf11b54263","Type":"ContainerDied","Data":"ecec43a9aa0edfd74c1bac14b8d026c06224210059018b414741ade266d698ca"} Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.011391 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.181544 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ssh-key-openstack-edpm-ipam\") pod \"3c6bade3-63d5-4342-8ded-feaf11b54263\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.181622 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c6bade3-63d5-4342-8ded-feaf11b54263-ovncontroller-config-0\") pod \"3c6bade3-63d5-4342-8ded-feaf11b54263\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.181768 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-inventory\") pod \"3c6bade3-63d5-4342-8ded-feaf11b54263\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.181861 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p6mh\" (UniqueName: \"kubernetes.io/projected/3c6bade3-63d5-4342-8ded-feaf11b54263-kube-api-access-6p6mh\") pod \"3c6bade3-63d5-4342-8ded-feaf11b54263\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.181952 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ovn-combined-ca-bundle\") pod \"3c6bade3-63d5-4342-8ded-feaf11b54263\" (UID: \"3c6bade3-63d5-4342-8ded-feaf11b54263\") " Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.192651 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3c6bade3-63d5-4342-8ded-feaf11b54263" (UID: "3c6bade3-63d5-4342-8ded-feaf11b54263"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.192705 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c6bade3-63d5-4342-8ded-feaf11b54263-kube-api-access-6p6mh" (OuterVolumeSpecName: "kube-api-access-6p6mh") pod "3c6bade3-63d5-4342-8ded-feaf11b54263" (UID: "3c6bade3-63d5-4342-8ded-feaf11b54263"). InnerVolumeSpecName "kube-api-access-6p6mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.212602 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-inventory" (OuterVolumeSpecName: "inventory") pod "3c6bade3-63d5-4342-8ded-feaf11b54263" (UID: "3c6bade3-63d5-4342-8ded-feaf11b54263"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.212835 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3c6bade3-63d5-4342-8ded-feaf11b54263" (UID: "3c6bade3-63d5-4342-8ded-feaf11b54263"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.224177 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c6bade3-63d5-4342-8ded-feaf11b54263-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "3c6bade3-63d5-4342-8ded-feaf11b54263" (UID: "3c6bade3-63d5-4342-8ded-feaf11b54263"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.283640 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.283821 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p6mh\" (UniqueName: \"kubernetes.io/projected/3c6bade3-63d5-4342-8ded-feaf11b54263-kube-api-access-6p6mh\") on node \"crc\" DevicePath \"\"" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.283913 4756 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.283966 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3c6bade3-63d5-4342-8ded-feaf11b54263-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.284022 4756 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3c6bade3-63d5-4342-8ded-feaf11b54263-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.664622 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" event={"ID":"3c6bade3-63d5-4342-8ded-feaf11b54263","Type":"ContainerDied","Data":"670374b188591e6166f19f8e3873145624ac3609ca4efc2a558846e046dfdb9d"} Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.664962 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="670374b188591e6166f19f8e3873145624ac3609ca4efc2a558846e046dfdb9d" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.664699 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jbd5l" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.732324 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f"] Feb 03 09:42:28 crc kubenswrapper[4756]: E0203 09:42:28.733199 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c6bade3-63d5-4342-8ded-feaf11b54263" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.733327 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c6bade3-63d5-4342-8ded-feaf11b54263" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.733635 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c6bade3-63d5-4342-8ded-feaf11b54263" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.734362 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.738008 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.738599 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.739395 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.739866 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.740001 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.742046 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.751532 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f"] Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.895585 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.895667 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7mpv\" (UniqueName: \"kubernetes.io/projected/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-kube-api-access-b7mpv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.895712 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.895741 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.895767 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.895857 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.997482 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.997596 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.997658 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7mpv\" (UniqueName: \"kubernetes.io/projected/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-kube-api-access-b7mpv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.997698 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.997727 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:28 crc kubenswrapper[4756]: I0203 09:42:28.997752 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.003539 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.003564 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.004469 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.011057 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.015008 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.023673 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7mpv\" (UniqueName: \"kubernetes.io/projected/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-kube-api-access-b7mpv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.054748 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.554934 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f"] Feb 03 09:42:29 crc kubenswrapper[4756]: I0203 09:42:29.674127 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" event={"ID":"cebe42d5-5283-4c1e-a2f2-70db6d3e2257","Type":"ContainerStarted","Data":"dec534ba189773b9f5b73dc31d26e5928a70bdd52e89e7623b7782c9b1ed7a2f"} Feb 03 09:42:30 crc kubenswrapper[4756]: I0203 09:42:30.683513 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" event={"ID":"cebe42d5-5283-4c1e-a2f2-70db6d3e2257","Type":"ContainerStarted","Data":"0021f6adf78b93ddc3461abcd93812494f286382c631fdf5bfacce17e4944cce"} Feb 03 09:42:30 crc kubenswrapper[4756]: I0203 09:42:30.710436 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" podStartSLOduration=2.532738773 podStartE2EDuration="2.710413105s" podCreationTimestamp="2026-02-03 09:42:28 +0000 UTC" firstStartedPulling="2026-02-03 09:42:29.561265276 +0000 UTC m=+1940.711732651" lastFinishedPulling="2026-02-03 09:42:29.738939608 +0000 UTC m=+1940.889406983" observedRunningTime="2026-02-03 09:42:30.702463797 +0000 UTC m=+1941.852931172" watchObservedRunningTime="2026-02-03 09:42:30.710413105 +0000 UTC m=+1941.860880520" Feb 03 09:42:43 crc kubenswrapper[4756]: I0203 09:42:43.566786 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:42:43 crc kubenswrapper[4756]: I0203 09:42:43.567287 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:43:13 crc kubenswrapper[4756]: I0203 09:43:13.565900 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:43:13 crc kubenswrapper[4756]: I0203 09:43:13.566494 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:43:14 crc kubenswrapper[4756]: I0203 09:43:14.054016 4756 generic.go:334] "Generic (PLEG): container finished" podID="cebe42d5-5283-4c1e-a2f2-70db6d3e2257" containerID="0021f6adf78b93ddc3461abcd93812494f286382c631fdf5bfacce17e4944cce" exitCode=0 Feb 03 09:43:14 crc kubenswrapper[4756]: I0203 09:43:14.054111 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" event={"ID":"cebe42d5-5283-4c1e-a2f2-70db6d3e2257","Type":"ContainerDied","Data":"0021f6adf78b93ddc3461abcd93812494f286382c631fdf5bfacce17e4944cce"} Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.452837 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.492101 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-inventory\") pod \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.492378 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-ssh-key-openstack-edpm-ipam\") pod \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.492433 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7mpv\" (UniqueName: \"kubernetes.io/projected/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-kube-api-access-b7mpv\") pod \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.492479 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-nova-metadata-neutron-config-0\") pod \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.492517 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-ovn-metadata-agent-neutron-config-0\") pod \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.492561 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-metadata-combined-ca-bundle\") pod \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\" (UID: \"cebe42d5-5283-4c1e-a2f2-70db6d3e2257\") " Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.498635 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cebe42d5-5283-4c1e-a2f2-70db6d3e2257" (UID: "cebe42d5-5283-4c1e-a2f2-70db6d3e2257"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.500190 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-kube-api-access-b7mpv" (OuterVolumeSpecName: "kube-api-access-b7mpv") pod "cebe42d5-5283-4c1e-a2f2-70db6d3e2257" (UID: "cebe42d5-5283-4c1e-a2f2-70db6d3e2257"). InnerVolumeSpecName "kube-api-access-b7mpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.519633 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-inventory" (OuterVolumeSpecName: "inventory") pod "cebe42d5-5283-4c1e-a2f2-70db6d3e2257" (UID: "cebe42d5-5283-4c1e-a2f2-70db6d3e2257"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.521376 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "cebe42d5-5283-4c1e-a2f2-70db6d3e2257" (UID: "cebe42d5-5283-4c1e-a2f2-70db6d3e2257"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.521599 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cebe42d5-5283-4c1e-a2f2-70db6d3e2257" (UID: "cebe42d5-5283-4c1e-a2f2-70db6d3e2257"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.523644 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "cebe42d5-5283-4c1e-a2f2-70db6d3e2257" (UID: "cebe42d5-5283-4c1e-a2f2-70db6d3e2257"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.595252 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.595296 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7mpv\" (UniqueName: \"kubernetes.io/projected/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-kube-api-access-b7mpv\") on node \"crc\" DevicePath \"\"" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.595305 4756 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.595315 4756 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.595329 4756 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:43:15 crc kubenswrapper[4756]: I0203 09:43:15.595338 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cebe42d5-5283-4c1e-a2f2-70db6d3e2257-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.073105 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" event={"ID":"cebe42d5-5283-4c1e-a2f2-70db6d3e2257","Type":"ContainerDied","Data":"dec534ba189773b9f5b73dc31d26e5928a70bdd52e89e7623b7782c9b1ed7a2f"} Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.073155 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dec534ba189773b9f5b73dc31d26e5928a70bdd52e89e7623b7782c9b1ed7a2f" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.073163 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.169359 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m"] Feb 03 09:43:16 crc kubenswrapper[4756]: E0203 09:43:16.169753 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cebe42d5-5283-4c1e-a2f2-70db6d3e2257" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.169778 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cebe42d5-5283-4c1e-a2f2-70db6d3e2257" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.169986 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="cebe42d5-5283-4c1e-a2f2-70db6d3e2257" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.179540 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.180597 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m"] Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.226418 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h56pt\" (UniqueName: \"kubernetes.io/projected/9419b87e-a027-4476-a95b-e4bc884d90e7-kube-api-access-h56pt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.226498 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.226535 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.226624 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.226652 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.226958 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.227164 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.227347 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.227485 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.227560 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.329712 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h56pt\" (UniqueName: \"kubernetes.io/projected/9419b87e-a027-4476-a95b-e4bc884d90e7-kube-api-access-h56pt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.329875 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.329951 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.330138 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.330213 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.335378 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.335898 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.336831 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.339835 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.347600 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h56pt\" (UniqueName: \"kubernetes.io/projected/9419b87e-a027-4476-a95b-e4bc884d90e7-kube-api-access-h56pt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:16 crc kubenswrapper[4756]: I0203 09:43:16.555299 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:43:17 crc kubenswrapper[4756]: I0203 09:43:17.085915 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m"] Feb 03 09:43:17 crc kubenswrapper[4756]: W0203 09:43:17.100719 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9419b87e_a027_4476_a95b_e4bc884d90e7.slice/crio-4737487d710d07bbabc91f6f380b0aafc1e79cb5aefdcc008c6f0059d3173ada WatchSource:0}: Error finding container 4737487d710d07bbabc91f6f380b0aafc1e79cb5aefdcc008c6f0059d3173ada: Status 404 returned error can't find the container with id 4737487d710d07bbabc91f6f380b0aafc1e79cb5aefdcc008c6f0059d3173ada Feb 03 09:43:18 crc kubenswrapper[4756]: I0203 09:43:18.090378 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" event={"ID":"9419b87e-a027-4476-a95b-e4bc884d90e7","Type":"ContainerStarted","Data":"4298a16a9a98a194248702967dc28a796821f97e1972a869bfb1a2c0c3c97996"} Feb 03 09:43:18 crc kubenswrapper[4756]: I0203 09:43:18.090783 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" event={"ID":"9419b87e-a027-4476-a95b-e4bc884d90e7","Type":"ContainerStarted","Data":"4737487d710d07bbabc91f6f380b0aafc1e79cb5aefdcc008c6f0059d3173ada"} Feb 03 09:43:18 crc kubenswrapper[4756]: I0203 09:43:18.140592 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" podStartSLOduration=1.97895417 podStartE2EDuration="2.140560813s" podCreationTimestamp="2026-02-03 09:43:16 +0000 UTC" firstStartedPulling="2026-02-03 09:43:17.103223075 +0000 UTC m=+1988.253690450" lastFinishedPulling="2026-02-03 09:43:17.264829718 +0000 UTC m=+1988.415297093" observedRunningTime="2026-02-03 09:43:18.114590544 +0000 UTC m=+1989.265057979" watchObservedRunningTime="2026-02-03 09:43:18.140560813 +0000 UTC m=+1989.291028208" Feb 03 09:43:43 crc kubenswrapper[4756]: I0203 09:43:43.565953 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:43:43 crc kubenswrapper[4756]: I0203 09:43:43.566652 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:43:43 crc kubenswrapper[4756]: I0203 09:43:43.566724 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:43:43 crc kubenswrapper[4756]: I0203 09:43:43.567662 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e6c9e3d9602222af7b0316088c411a3230434432a9c9562494d24bc001e56b1"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:43:43 crc kubenswrapper[4756]: I0203 09:43:43.567744 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://6e6c9e3d9602222af7b0316088c411a3230434432a9c9562494d24bc001e56b1" gracePeriod=600 Feb 03 09:43:44 crc kubenswrapper[4756]: I0203 09:43:44.352139 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="6e6c9e3d9602222af7b0316088c411a3230434432a9c9562494d24bc001e56b1" exitCode=0 Feb 03 09:43:44 crc kubenswrapper[4756]: I0203 09:43:44.352577 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"6e6c9e3d9602222af7b0316088c411a3230434432a9c9562494d24bc001e56b1"} Feb 03 09:43:44 crc kubenswrapper[4756]: I0203 09:43:44.352624 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945"} Feb 03 09:43:44 crc kubenswrapper[4756]: I0203 09:43:44.352650 4756 scope.go:117] "RemoveContainer" containerID="c9dfc831708010ffdd8cb355aed2cbfbfd0bd4422689a6bea4268598cc427437" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.018143 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j9qm4"] Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.020518 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.032624 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9qm4"] Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.084959 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6gkm\" (UniqueName: \"kubernetes.io/projected/7b508806-6330-407b-8a6f-60da6252041d-kube-api-access-c6gkm\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.085102 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-catalog-content\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.085242 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-utilities\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.186799 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-catalog-content\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.186919 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-utilities\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.186950 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6gkm\" (UniqueName: \"kubernetes.io/projected/7b508806-6330-407b-8a6f-60da6252041d-kube-api-access-c6gkm\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.187384 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-utilities\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.187377 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-catalog-content\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.206889 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6gkm\" (UniqueName: \"kubernetes.io/projected/7b508806-6330-407b-8a6f-60da6252041d-kube-api-access-c6gkm\") pod \"redhat-operators-j9qm4\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.345614 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:29 crc kubenswrapper[4756]: I0203 09:44:29.790799 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9qm4"] Feb 03 09:44:29 crc kubenswrapper[4756]: W0203 09:44:29.797164 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b508806_6330_407b_8a6f_60da6252041d.slice/crio-adada47e23f9c833ab0f785d2572fb9f0c2729e553c0bccc60eeac3e8c42b861 WatchSource:0}: Error finding container adada47e23f9c833ab0f785d2572fb9f0c2729e553c0bccc60eeac3e8c42b861: Status 404 returned error can't find the container with id adada47e23f9c833ab0f785d2572fb9f0c2729e553c0bccc60eeac3e8c42b861 Feb 03 09:44:30 crc kubenswrapper[4756]: I0203 09:44:30.770327 4756 generic.go:334] "Generic (PLEG): container finished" podID="7b508806-6330-407b-8a6f-60da6252041d" containerID="118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266" exitCode=0 Feb 03 09:44:30 crc kubenswrapper[4756]: I0203 09:44:30.770381 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qm4" event={"ID":"7b508806-6330-407b-8a6f-60da6252041d","Type":"ContainerDied","Data":"118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266"} Feb 03 09:44:30 crc kubenswrapper[4756]: I0203 09:44:30.770842 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qm4" event={"ID":"7b508806-6330-407b-8a6f-60da6252041d","Type":"ContainerStarted","Data":"adada47e23f9c833ab0f785d2572fb9f0c2729e553c0bccc60eeac3e8c42b861"} Feb 03 09:44:30 crc kubenswrapper[4756]: I0203 09:44:30.772585 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 09:44:31 crc kubenswrapper[4756]: I0203 09:44:31.780830 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qm4" event={"ID":"7b508806-6330-407b-8a6f-60da6252041d","Type":"ContainerStarted","Data":"867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870"} Feb 03 09:44:32 crc kubenswrapper[4756]: I0203 09:44:32.794278 4756 generic.go:334] "Generic (PLEG): container finished" podID="7b508806-6330-407b-8a6f-60da6252041d" containerID="867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870" exitCode=0 Feb 03 09:44:32 crc kubenswrapper[4756]: I0203 09:44:32.794351 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qm4" event={"ID":"7b508806-6330-407b-8a6f-60da6252041d","Type":"ContainerDied","Data":"867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870"} Feb 03 09:44:33 crc kubenswrapper[4756]: I0203 09:44:33.804135 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qm4" event={"ID":"7b508806-6330-407b-8a6f-60da6252041d","Type":"ContainerStarted","Data":"159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780"} Feb 03 09:44:33 crc kubenswrapper[4756]: I0203 09:44:33.826286 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j9qm4" podStartSLOduration=3.4035373939999998 podStartE2EDuration="5.826267134s" podCreationTimestamp="2026-02-03 09:44:28 +0000 UTC" firstStartedPulling="2026-02-03 09:44:30.77229519 +0000 UTC m=+2061.922762565" lastFinishedPulling="2026-02-03 09:44:33.19502491 +0000 UTC m=+2064.345492305" observedRunningTime="2026-02-03 09:44:33.820709391 +0000 UTC m=+2064.971176776" watchObservedRunningTime="2026-02-03 09:44:33.826267134 +0000 UTC m=+2064.976734519" Feb 03 09:44:39 crc kubenswrapper[4756]: I0203 09:44:39.346029 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:39 crc kubenswrapper[4756]: I0203 09:44:39.346359 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:40 crc kubenswrapper[4756]: I0203 09:44:40.387656 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j9qm4" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="registry-server" probeResult="failure" output=< Feb 03 09:44:40 crc kubenswrapper[4756]: timeout: failed to connect service ":50051" within 1s Feb 03 09:44:40 crc kubenswrapper[4756]: > Feb 03 09:44:49 crc kubenswrapper[4756]: I0203 09:44:49.396905 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:49 crc kubenswrapper[4756]: I0203 09:44:49.451765 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:49 crc kubenswrapper[4756]: I0203 09:44:49.641154 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9qm4"] Feb 03 09:44:50 crc kubenswrapper[4756]: I0203 09:44:50.948106 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j9qm4" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="registry-server" containerID="cri-o://159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780" gracePeriod=2 Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.385158 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.542272 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-utilities\") pod \"7b508806-6330-407b-8a6f-60da6252041d\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.542462 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-catalog-content\") pod \"7b508806-6330-407b-8a6f-60da6252041d\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.542546 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6gkm\" (UniqueName: \"kubernetes.io/projected/7b508806-6330-407b-8a6f-60da6252041d-kube-api-access-c6gkm\") pod \"7b508806-6330-407b-8a6f-60da6252041d\" (UID: \"7b508806-6330-407b-8a6f-60da6252041d\") " Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.543109 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-utilities" (OuterVolumeSpecName: "utilities") pod "7b508806-6330-407b-8a6f-60da6252041d" (UID: "7b508806-6330-407b-8a6f-60da6252041d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.544468 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.550590 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b508806-6330-407b-8a6f-60da6252041d-kube-api-access-c6gkm" (OuterVolumeSpecName: "kube-api-access-c6gkm") pod "7b508806-6330-407b-8a6f-60da6252041d" (UID: "7b508806-6330-407b-8a6f-60da6252041d"). InnerVolumeSpecName "kube-api-access-c6gkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.646082 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b508806-6330-407b-8a6f-60da6252041d" (UID: "7b508806-6330-407b-8a6f-60da6252041d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.646895 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b508806-6330-407b-8a6f-60da6252041d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.646912 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6gkm\" (UniqueName: \"kubernetes.io/projected/7b508806-6330-407b-8a6f-60da6252041d-kube-api-access-c6gkm\") on node \"crc\" DevicePath \"\"" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.967127 4756 generic.go:334] "Generic (PLEG): container finished" podID="7b508806-6330-407b-8a6f-60da6252041d" containerID="159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780" exitCode=0 Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.967168 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qm4" event={"ID":"7b508806-6330-407b-8a6f-60da6252041d","Type":"ContainerDied","Data":"159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780"} Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.967180 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9qm4" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.967195 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qm4" event={"ID":"7b508806-6330-407b-8a6f-60da6252041d","Type":"ContainerDied","Data":"adada47e23f9c833ab0f785d2572fb9f0c2729e553c0bccc60eeac3e8c42b861"} Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.967215 4756 scope.go:117] "RemoveContainer" containerID="159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780" Feb 03 09:44:51 crc kubenswrapper[4756]: I0203 09:44:51.999073 4756 scope.go:117] "RemoveContainer" containerID="867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870" Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.008099 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9qm4"] Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.016629 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j9qm4"] Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.020138 4756 scope.go:117] "RemoveContainer" containerID="118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266" Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.061985 4756 scope.go:117] "RemoveContainer" containerID="159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780" Feb 03 09:44:52 crc kubenswrapper[4756]: E0203 09:44:52.062490 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780\": container with ID starting with 159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780 not found: ID does not exist" containerID="159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780" Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.062543 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780"} err="failed to get container status \"159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780\": rpc error: code = NotFound desc = could not find container \"159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780\": container with ID starting with 159c696c03232b6dcf4eb1447e11ffe12f65fb8d46e6f41a30ff56327fbf8780 not found: ID does not exist" Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.062574 4756 scope.go:117] "RemoveContainer" containerID="867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870" Feb 03 09:44:52 crc kubenswrapper[4756]: E0203 09:44:52.063067 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870\": container with ID starting with 867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870 not found: ID does not exist" containerID="867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870" Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.063182 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870"} err="failed to get container status \"867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870\": rpc error: code = NotFound desc = could not find container \"867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870\": container with ID starting with 867ff6ba232167e8aaa2a4c76e53032e32ef63a4bd40beb03ae2804924f01870 not found: ID does not exist" Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.063326 4756 scope.go:117] "RemoveContainer" containerID="118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266" Feb 03 09:44:52 crc kubenswrapper[4756]: E0203 09:44:52.063884 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266\": container with ID starting with 118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266 not found: ID does not exist" containerID="118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266" Feb 03 09:44:52 crc kubenswrapper[4756]: I0203 09:44:52.063911 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266"} err="failed to get container status \"118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266\": rpc error: code = NotFound desc = could not find container \"118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266\": container with ID starting with 118132714a644e1a92836b73007d3db34a267c8c5c5baaa4b0b0043752c2e266 not found: ID does not exist" Feb 03 09:44:53 crc kubenswrapper[4756]: I0203 09:44:53.625040 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b508806-6330-407b-8a6f-60da6252041d" path="/var/lib/kubelet/pods/7b508806-6330-407b-8a6f-60da6252041d/volumes" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.147182 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4"] Feb 03 09:45:00 crc kubenswrapper[4756]: E0203 09:45:00.148346 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="extract-content" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.148364 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="extract-content" Feb 03 09:45:00 crc kubenswrapper[4756]: E0203 09:45:00.148404 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="extract-utilities" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.148414 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="extract-utilities" Feb 03 09:45:00 crc kubenswrapper[4756]: E0203 09:45:00.148436 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="registry-server" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.148463 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="registry-server" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.148690 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b508806-6330-407b-8a6f-60da6252041d" containerName="registry-server" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.149620 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.156657 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.159257 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.162751 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4"] Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.309400 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e432f22-68e6-4ec1-9825-de27403d0eeb-secret-volume\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.310184 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e432f22-68e6-4ec1-9825-de27403d0eeb-config-volume\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.310396 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knr5c\" (UniqueName: \"kubernetes.io/projected/6e432f22-68e6-4ec1-9825-de27403d0eeb-kube-api-access-knr5c\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.412325 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knr5c\" (UniqueName: \"kubernetes.io/projected/6e432f22-68e6-4ec1-9825-de27403d0eeb-kube-api-access-knr5c\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.412477 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e432f22-68e6-4ec1-9825-de27403d0eeb-secret-volume\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.412509 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e432f22-68e6-4ec1-9825-de27403d0eeb-config-volume\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.413612 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e432f22-68e6-4ec1-9825-de27403d0eeb-config-volume\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.426420 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e432f22-68e6-4ec1-9825-de27403d0eeb-secret-volume\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.431033 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knr5c\" (UniqueName: \"kubernetes.io/projected/6e432f22-68e6-4ec1-9825-de27403d0eeb-kube-api-access-knr5c\") pod \"collect-profiles-29501865-s9cc4\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.483168 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:00 crc kubenswrapper[4756]: I0203 09:45:00.950120 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4"] Feb 03 09:45:01 crc kubenswrapper[4756]: I0203 09:45:01.047590 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" event={"ID":"6e432f22-68e6-4ec1-9825-de27403d0eeb","Type":"ContainerStarted","Data":"a244e2e0da7be1ee7f4f7cba7003279666e0b22c1535c812355cf0b5ea999549"} Feb 03 09:45:02 crc kubenswrapper[4756]: I0203 09:45:02.061423 4756 generic.go:334] "Generic (PLEG): container finished" podID="6e432f22-68e6-4ec1-9825-de27403d0eeb" containerID="f0ffbe5cc2548ac01ecf6d4cd278ecd0b9f5834b3e3f7fcdede8acadbc057359" exitCode=0 Feb 03 09:45:02 crc kubenswrapper[4756]: I0203 09:45:02.061562 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" event={"ID":"6e432f22-68e6-4ec1-9825-de27403d0eeb","Type":"ContainerDied","Data":"f0ffbe5cc2548ac01ecf6d4cd278ecd0b9f5834b3e3f7fcdede8acadbc057359"} Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.399862 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.572486 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knr5c\" (UniqueName: \"kubernetes.io/projected/6e432f22-68e6-4ec1-9825-de27403d0eeb-kube-api-access-knr5c\") pod \"6e432f22-68e6-4ec1-9825-de27403d0eeb\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.572570 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e432f22-68e6-4ec1-9825-de27403d0eeb-secret-volume\") pod \"6e432f22-68e6-4ec1-9825-de27403d0eeb\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.572726 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e432f22-68e6-4ec1-9825-de27403d0eeb-config-volume\") pod \"6e432f22-68e6-4ec1-9825-de27403d0eeb\" (UID: \"6e432f22-68e6-4ec1-9825-de27403d0eeb\") " Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.573435 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e432f22-68e6-4ec1-9825-de27403d0eeb-config-volume" (OuterVolumeSpecName: "config-volume") pod "6e432f22-68e6-4ec1-9825-de27403d0eeb" (UID: "6e432f22-68e6-4ec1-9825-de27403d0eeb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.578229 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e432f22-68e6-4ec1-9825-de27403d0eeb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6e432f22-68e6-4ec1-9825-de27403d0eeb" (UID: "6e432f22-68e6-4ec1-9825-de27403d0eeb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.578468 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e432f22-68e6-4ec1-9825-de27403d0eeb-kube-api-access-knr5c" (OuterVolumeSpecName: "kube-api-access-knr5c") pod "6e432f22-68e6-4ec1-9825-de27403d0eeb" (UID: "6e432f22-68e6-4ec1-9825-de27403d0eeb"). InnerVolumeSpecName "kube-api-access-knr5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.674543 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knr5c\" (UniqueName: \"kubernetes.io/projected/6e432f22-68e6-4ec1-9825-de27403d0eeb-kube-api-access-knr5c\") on node \"crc\" DevicePath \"\"" Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.674578 4756 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e432f22-68e6-4ec1-9825-de27403d0eeb-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:45:03 crc kubenswrapper[4756]: I0203 09:45:03.674591 4756 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e432f22-68e6-4ec1-9825-de27403d0eeb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 03 09:45:04 crc kubenswrapper[4756]: I0203 09:45:04.084363 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" event={"ID":"6e432f22-68e6-4ec1-9825-de27403d0eeb","Type":"ContainerDied","Data":"a244e2e0da7be1ee7f4f7cba7003279666e0b22c1535c812355cf0b5ea999549"} Feb 03 09:45:04 crc kubenswrapper[4756]: I0203 09:45:04.084429 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a244e2e0da7be1ee7f4f7cba7003279666e0b22c1535c812355cf0b5ea999549" Feb 03 09:45:04 crc kubenswrapper[4756]: I0203 09:45:04.084575 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501865-s9cc4" Feb 03 09:45:04 crc kubenswrapper[4756]: I0203 09:45:04.476077 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z"] Feb 03 09:45:04 crc kubenswrapper[4756]: I0203 09:45:04.486232 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501820-8bb2z"] Feb 03 09:45:05 crc kubenswrapper[4756]: I0203 09:45:05.628491 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e742c048-9564-4b75-986b-a34639494020" path="/var/lib/kubelet/pods/e742c048-9564-4b75-986b-a34639494020/volumes" Feb 03 09:45:31 crc kubenswrapper[4756]: I0203 09:45:31.411907 4756 scope.go:117] "RemoveContainer" containerID="ff07b217cc643322909bb9ab3d7a77b96673d00b9ba816c5d024d1b3056c44a2" Feb 03 09:45:43 crc kubenswrapper[4756]: I0203 09:45:43.566737 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:45:43 crc kubenswrapper[4756]: I0203 09:45:43.567711 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:46:13 crc kubenswrapper[4756]: I0203 09:46:13.566667 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:46:13 crc kubenswrapper[4756]: I0203 09:46:13.569119 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:46:22 crc kubenswrapper[4756]: I0203 09:46:22.946226 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wcbcv"] Feb 03 09:46:22 crc kubenswrapper[4756]: E0203 09:46:22.947732 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e432f22-68e6-4ec1-9825-de27403d0eeb" containerName="collect-profiles" Feb 03 09:46:22 crc kubenswrapper[4756]: I0203 09:46:22.947751 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e432f22-68e6-4ec1-9825-de27403d0eeb" containerName="collect-profiles" Feb 03 09:46:22 crc kubenswrapper[4756]: I0203 09:46:22.948047 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e432f22-68e6-4ec1-9825-de27403d0eeb" containerName="collect-profiles" Feb 03 09:46:22 crc kubenswrapper[4756]: I0203 09:46:22.965706 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wcbcv"] Feb 03 09:46:22 crc kubenswrapper[4756]: I0203 09:46:22.965853 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.043377 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf72p\" (UniqueName: \"kubernetes.io/projected/54b83190-4235-4c19-8f43-2c641bf1595d-kube-api-access-nf72p\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.043420 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-catalog-content\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.043564 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-utilities\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.144696 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf72p\" (UniqueName: \"kubernetes.io/projected/54b83190-4235-4c19-8f43-2c641bf1595d-kube-api-access-nf72p\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.144744 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-catalog-content\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.144842 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-utilities\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.145505 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-utilities\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.145507 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-catalog-content\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.163786 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf72p\" (UniqueName: \"kubernetes.io/projected/54b83190-4235-4c19-8f43-2c641bf1595d-kube-api-access-nf72p\") pod \"certified-operators-wcbcv\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.290932 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:23 crc kubenswrapper[4756]: I0203 09:46:23.823782 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wcbcv"] Feb 03 09:46:24 crc kubenswrapper[4756]: I0203 09:46:24.756342 4756 generic.go:334] "Generic (PLEG): container finished" podID="54b83190-4235-4c19-8f43-2c641bf1595d" containerID="56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8" exitCode=0 Feb 03 09:46:24 crc kubenswrapper[4756]: I0203 09:46:24.756577 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcbcv" event={"ID":"54b83190-4235-4c19-8f43-2c641bf1595d","Type":"ContainerDied","Data":"56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8"} Feb 03 09:46:24 crc kubenswrapper[4756]: I0203 09:46:24.756855 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcbcv" event={"ID":"54b83190-4235-4c19-8f43-2c641bf1595d","Type":"ContainerStarted","Data":"96d1b7e431e81183c44fe2266c24e5e55df5b7c9059b2ea0cda78759b0cb7701"} Feb 03 09:46:25 crc kubenswrapper[4756]: I0203 09:46:25.766561 4756 generic.go:334] "Generic (PLEG): container finished" podID="54b83190-4235-4c19-8f43-2c641bf1595d" containerID="6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d" exitCode=0 Feb 03 09:46:25 crc kubenswrapper[4756]: I0203 09:46:25.766639 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcbcv" event={"ID":"54b83190-4235-4c19-8f43-2c641bf1595d","Type":"ContainerDied","Data":"6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d"} Feb 03 09:46:26 crc kubenswrapper[4756]: I0203 09:46:26.778946 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcbcv" event={"ID":"54b83190-4235-4c19-8f43-2c641bf1595d","Type":"ContainerStarted","Data":"f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358"} Feb 03 09:46:26 crc kubenswrapper[4756]: I0203 09:46:26.801974 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wcbcv" podStartSLOduration=3.380585087 podStartE2EDuration="4.801955075s" podCreationTimestamp="2026-02-03 09:46:22 +0000 UTC" firstStartedPulling="2026-02-03 09:46:24.758842519 +0000 UTC m=+2175.909309904" lastFinishedPulling="2026-02-03 09:46:26.180212517 +0000 UTC m=+2177.330679892" observedRunningTime="2026-02-03 09:46:26.795810194 +0000 UTC m=+2177.946277569" watchObservedRunningTime="2026-02-03 09:46:26.801955075 +0000 UTC m=+2177.952422450" Feb 03 09:46:33 crc kubenswrapper[4756]: I0203 09:46:33.291047 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:33 crc kubenswrapper[4756]: I0203 09:46:33.292611 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:33 crc kubenswrapper[4756]: I0203 09:46:33.336669 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:33 crc kubenswrapper[4756]: I0203 09:46:33.913756 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:34 crc kubenswrapper[4756]: I0203 09:46:34.744169 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wcbcv"] Feb 03 09:46:35 crc kubenswrapper[4756]: I0203 09:46:35.857773 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wcbcv" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" containerName="registry-server" containerID="cri-o://f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358" gracePeriod=2 Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.313766 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.401992 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-utilities\") pod \"54b83190-4235-4c19-8f43-2c641bf1595d\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.402245 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf72p\" (UniqueName: \"kubernetes.io/projected/54b83190-4235-4c19-8f43-2c641bf1595d-kube-api-access-nf72p\") pod \"54b83190-4235-4c19-8f43-2c641bf1595d\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.402293 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-catalog-content\") pod \"54b83190-4235-4c19-8f43-2c641bf1595d\" (UID: \"54b83190-4235-4c19-8f43-2c641bf1595d\") " Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.402884 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-utilities" (OuterVolumeSpecName: "utilities") pod "54b83190-4235-4c19-8f43-2c641bf1595d" (UID: "54b83190-4235-4c19-8f43-2c641bf1595d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.409833 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54b83190-4235-4c19-8f43-2c641bf1595d-kube-api-access-nf72p" (OuterVolumeSpecName: "kube-api-access-nf72p") pod "54b83190-4235-4c19-8f43-2c641bf1595d" (UID: "54b83190-4235-4c19-8f43-2c641bf1595d"). InnerVolumeSpecName "kube-api-access-nf72p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.504526 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf72p\" (UniqueName: \"kubernetes.io/projected/54b83190-4235-4c19-8f43-2c641bf1595d-kube-api-access-nf72p\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.504564 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.524276 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54b83190-4235-4c19-8f43-2c641bf1595d" (UID: "54b83190-4235-4c19-8f43-2c641bf1595d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.607395 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54b83190-4235-4c19-8f43-2c641bf1595d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.867999 4756 generic.go:334] "Generic (PLEG): container finished" podID="54b83190-4235-4c19-8f43-2c641bf1595d" containerID="f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358" exitCode=0 Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.868052 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcbcv" event={"ID":"54b83190-4235-4c19-8f43-2c641bf1595d","Type":"ContainerDied","Data":"f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358"} Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.868076 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcbcv" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.868102 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcbcv" event={"ID":"54b83190-4235-4c19-8f43-2c641bf1595d","Type":"ContainerDied","Data":"96d1b7e431e81183c44fe2266c24e5e55df5b7c9059b2ea0cda78759b0cb7701"} Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.868124 4756 scope.go:117] "RemoveContainer" containerID="f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.887716 4756 scope.go:117] "RemoveContainer" containerID="6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.903133 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wcbcv"] Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.910533 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wcbcv"] Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.926208 4756 scope.go:117] "RemoveContainer" containerID="56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.962724 4756 scope.go:117] "RemoveContainer" containerID="f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358" Feb 03 09:46:36 crc kubenswrapper[4756]: E0203 09:46:36.963976 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358\": container with ID starting with f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358 not found: ID does not exist" containerID="f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.964051 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358"} err="failed to get container status \"f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358\": rpc error: code = NotFound desc = could not find container \"f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358\": container with ID starting with f54a9d6e8dcf58744435a0354209b8e46ca8dada8bf9b6823ca0b0b536701358 not found: ID does not exist" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.964094 4756 scope.go:117] "RemoveContainer" containerID="6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d" Feb 03 09:46:36 crc kubenswrapper[4756]: E0203 09:46:36.964533 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d\": container with ID starting with 6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d not found: ID does not exist" containerID="6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.964575 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d"} err="failed to get container status \"6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d\": rpc error: code = NotFound desc = could not find container \"6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d\": container with ID starting with 6107a11628690888fc4e81957c11d7a5bfd17745f052bd28adb58aaf1f67de2d not found: ID does not exist" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.964602 4756 scope.go:117] "RemoveContainer" containerID="56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8" Feb 03 09:46:36 crc kubenswrapper[4756]: E0203 09:46:36.965291 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8\": container with ID starting with 56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8 not found: ID does not exist" containerID="56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8" Feb 03 09:46:36 crc kubenswrapper[4756]: I0203 09:46:36.965340 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8"} err="failed to get container status \"56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8\": rpc error: code = NotFound desc = could not find container \"56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8\": container with ID starting with 56f5f6b8a83dd98cf1da323ba7c53cbdd7fb7d63c691b7a928865e24723ee0e8 not found: ID does not exist" Feb 03 09:46:37 crc kubenswrapper[4756]: I0203 09:46:37.625377 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" path="/var/lib/kubelet/pods/54b83190-4235-4c19-8f43-2c641bf1595d/volumes" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.421148 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rpp52"] Feb 03 09:46:41 crc kubenswrapper[4756]: E0203 09:46:41.424983 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" containerName="extract-utilities" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.425013 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" containerName="extract-utilities" Feb 03 09:46:41 crc kubenswrapper[4756]: E0203 09:46:41.425064 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" containerName="extract-content" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.425072 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" containerName="extract-content" Feb 03 09:46:41 crc kubenswrapper[4756]: E0203 09:46:41.425083 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" containerName="registry-server" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.425090 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" containerName="registry-server" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.425301 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="54b83190-4235-4c19-8f43-2c641bf1595d" containerName="registry-server" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.426948 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.445000 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpp52"] Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.598946 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx4wm\" (UniqueName: \"kubernetes.io/projected/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-kube-api-access-wx4wm\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.599011 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-utilities\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.599196 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-catalog-content\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.700748 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx4wm\" (UniqueName: \"kubernetes.io/projected/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-kube-api-access-wx4wm\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.700847 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-utilities\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.700931 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-catalog-content\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.701409 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-utilities\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.701491 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-catalog-content\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.722339 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx4wm\" (UniqueName: \"kubernetes.io/projected/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-kube-api-access-wx4wm\") pod \"community-operators-rpp52\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:41 crc kubenswrapper[4756]: I0203 09:46:41.779150 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:42 crc kubenswrapper[4756]: I0203 09:46:42.339104 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpp52"] Feb 03 09:46:42 crc kubenswrapper[4756]: I0203 09:46:42.919785 4756 generic.go:334] "Generic (PLEG): container finished" podID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerID="bc40ba2ba06cfc3d04dddc1acb214338ec886363973802a3ce35bfbba116e7b3" exitCode=0 Feb 03 09:46:42 crc kubenswrapper[4756]: I0203 09:46:42.919849 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp52" event={"ID":"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126","Type":"ContainerDied","Data":"bc40ba2ba06cfc3d04dddc1acb214338ec886363973802a3ce35bfbba116e7b3"} Feb 03 09:46:42 crc kubenswrapper[4756]: I0203 09:46:42.920096 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp52" event={"ID":"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126","Type":"ContainerStarted","Data":"dc16bc5541c2cc7b8ceec0cc6f00ce24bf2e7f5caa12434d6c3bae3a0eabfc84"} Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.569179 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.569611 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.569667 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.570756 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.570816 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" gracePeriod=600 Feb 03 09:46:43 crc kubenswrapper[4756]: E0203 09:46:43.738347 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.929731 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" exitCode=0 Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.929794 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945"} Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.929832 4756 scope.go:117] "RemoveContainer" containerID="6e6c9e3d9602222af7b0316088c411a3230434432a9c9562494d24bc001e56b1" Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.930518 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:46:43 crc kubenswrapper[4756]: E0203 09:46:43.930844 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:46:43 crc kubenswrapper[4756]: I0203 09:46:43.933896 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp52" event={"ID":"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126","Type":"ContainerStarted","Data":"731c05051c9ad49546659919b8e3dc3bb37e4e001915a500d9fecc2d8ed74dae"} Feb 03 09:46:44 crc kubenswrapper[4756]: I0203 09:46:44.944937 4756 generic.go:334] "Generic (PLEG): container finished" podID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerID="731c05051c9ad49546659919b8e3dc3bb37e4e001915a500d9fecc2d8ed74dae" exitCode=0 Feb 03 09:46:44 crc kubenswrapper[4756]: I0203 09:46:44.945147 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp52" event={"ID":"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126","Type":"ContainerDied","Data":"731c05051c9ad49546659919b8e3dc3bb37e4e001915a500d9fecc2d8ed74dae"} Feb 03 09:46:45 crc kubenswrapper[4756]: I0203 09:46:45.956033 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp52" event={"ID":"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126","Type":"ContainerStarted","Data":"869021e66840c8836ab031f7fa86742657bcdd7bd29f35b425a1ee2cf95600ff"} Feb 03 09:46:45 crc kubenswrapper[4756]: I0203 09:46:45.978843 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rpp52" podStartSLOduration=2.561775281 podStartE2EDuration="4.978818138s" podCreationTimestamp="2026-02-03 09:46:41 +0000 UTC" firstStartedPulling="2026-02-03 09:46:42.922595582 +0000 UTC m=+2194.073062957" lastFinishedPulling="2026-02-03 09:46:45.339638439 +0000 UTC m=+2196.490105814" observedRunningTime="2026-02-03 09:46:45.972950095 +0000 UTC m=+2197.123417480" watchObservedRunningTime="2026-02-03 09:46:45.978818138 +0000 UTC m=+2197.129285523" Feb 03 09:46:47 crc kubenswrapper[4756]: I0203 09:46:47.975152 4756 generic.go:334] "Generic (PLEG): container finished" podID="9419b87e-a027-4476-a95b-e4bc884d90e7" containerID="4298a16a9a98a194248702967dc28a796821f97e1972a869bfb1a2c0c3c97996" exitCode=0 Feb 03 09:46:47 crc kubenswrapper[4756]: I0203 09:46:47.975292 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" event={"ID":"9419b87e-a027-4476-a95b-e4bc884d90e7","Type":"ContainerDied","Data":"4298a16a9a98a194248702967dc28a796821f97e1972a869bfb1a2c0c3c97996"} Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.505234 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.645542 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-ssh-key-openstack-edpm-ipam\") pod \"9419b87e-a027-4476-a95b-e4bc884d90e7\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.646190 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-combined-ca-bundle\") pod \"9419b87e-a027-4476-a95b-e4bc884d90e7\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.646304 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-inventory\") pod \"9419b87e-a027-4476-a95b-e4bc884d90e7\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.646425 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-secret-0\") pod \"9419b87e-a027-4476-a95b-e4bc884d90e7\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.646603 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h56pt\" (UniqueName: \"kubernetes.io/projected/9419b87e-a027-4476-a95b-e4bc884d90e7-kube-api-access-h56pt\") pod \"9419b87e-a027-4476-a95b-e4bc884d90e7\" (UID: \"9419b87e-a027-4476-a95b-e4bc884d90e7\") " Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.651496 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9419b87e-a027-4476-a95b-e4bc884d90e7-kube-api-access-h56pt" (OuterVolumeSpecName: "kube-api-access-h56pt") pod "9419b87e-a027-4476-a95b-e4bc884d90e7" (UID: "9419b87e-a027-4476-a95b-e4bc884d90e7"). InnerVolumeSpecName "kube-api-access-h56pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.652695 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "9419b87e-a027-4476-a95b-e4bc884d90e7" (UID: "9419b87e-a027-4476-a95b-e4bc884d90e7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.673006 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9419b87e-a027-4476-a95b-e4bc884d90e7" (UID: "9419b87e-a027-4476-a95b-e4bc884d90e7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.674480 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "9419b87e-a027-4476-a95b-e4bc884d90e7" (UID: "9419b87e-a027-4476-a95b-e4bc884d90e7"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.684980 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-inventory" (OuterVolumeSpecName: "inventory") pod "9419b87e-a027-4476-a95b-e4bc884d90e7" (UID: "9419b87e-a027-4476-a95b-e4bc884d90e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.749812 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h56pt\" (UniqueName: \"kubernetes.io/projected/9419b87e-a027-4476-a95b-e4bc884d90e7-kube-api-access-h56pt\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.750336 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.750350 4756 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.750359 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.750370 4756 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9419b87e-a027-4476-a95b-e4bc884d90e7-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.999326 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" event={"ID":"9419b87e-a027-4476-a95b-e4bc884d90e7","Type":"ContainerDied","Data":"4737487d710d07bbabc91f6f380b0aafc1e79cb5aefdcc008c6f0059d3173ada"} Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.999498 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m" Feb 03 09:46:49 crc kubenswrapper[4756]: I0203 09:46:49.999505 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4737487d710d07bbabc91f6f380b0aafc1e79cb5aefdcc008c6f0059d3173ada" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.116258 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp"] Feb 03 09:46:50 crc kubenswrapper[4756]: E0203 09:46:50.116700 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9419b87e-a027-4476-a95b-e4bc884d90e7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.116719 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="9419b87e-a027-4476-a95b-e4bc884d90e7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.116886 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="9419b87e-a027-4476-a95b-e4bc884d90e7" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.117538 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.123250 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.123304 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.123432 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.123473 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.123264 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.123522 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.130329 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.130465 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp"] Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.260645 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d72ck\" (UniqueName: \"kubernetes.io/projected/0573a518-65e1-4717-a24e-285e30cad247-kube-api-access-d72ck\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.260712 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.260785 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.260807 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.260834 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0573a518-65e1-4717-a24e-285e30cad247-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.260894 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.261012 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.261094 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.261147 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.362805 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.363071 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.363094 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.363111 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0573a518-65e1-4717-a24e-285e30cad247-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.363137 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.363167 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.363193 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.363216 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.364104 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d72ck\" (UniqueName: \"kubernetes.io/projected/0573a518-65e1-4717-a24e-285e30cad247-kube-api-access-d72ck\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.364113 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0573a518-65e1-4717-a24e-285e30cad247-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.367940 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.367986 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.368589 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.369259 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.371881 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.378188 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.378784 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.379833 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d72ck\" (UniqueName: \"kubernetes.io/projected/0573a518-65e1-4717-a24e-285e30cad247-kube-api-access-d72ck\") pod \"nova-edpm-deployment-openstack-edpm-ipam-k5dsp\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.440747 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:46:50 crc kubenswrapper[4756]: I0203 09:46:50.966760 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp"] Feb 03 09:46:50 crc kubenswrapper[4756]: W0203 09:46:50.979487 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0573a518_65e1_4717_a24e_285e30cad247.slice/crio-39a2447e07be04d4aed03fc6d3d096fb84772955dd47a596a40b4b5e34b3803f WatchSource:0}: Error finding container 39a2447e07be04d4aed03fc6d3d096fb84772955dd47a596a40b4b5e34b3803f: Status 404 returned error can't find the container with id 39a2447e07be04d4aed03fc6d3d096fb84772955dd47a596a40b4b5e34b3803f Feb 03 09:46:51 crc kubenswrapper[4756]: I0203 09:46:51.011274 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" event={"ID":"0573a518-65e1-4717-a24e-285e30cad247","Type":"ContainerStarted","Data":"39a2447e07be04d4aed03fc6d3d096fb84772955dd47a596a40b4b5e34b3803f"} Feb 03 09:46:51 crc kubenswrapper[4756]: I0203 09:46:51.780200 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:51 crc kubenswrapper[4756]: I0203 09:46:51.780467 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:51 crc kubenswrapper[4756]: I0203 09:46:51.833778 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:52 crc kubenswrapper[4756]: I0203 09:46:52.021662 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" event={"ID":"0573a518-65e1-4717-a24e-285e30cad247","Type":"ContainerStarted","Data":"d52768fa1e2734e1c4d8d17e999a0ba1e0f7a38feae43da55a29ab607ed64dfa"} Feb 03 09:46:52 crc kubenswrapper[4756]: I0203 09:46:52.052956 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" podStartSLOduration=1.8699787209999998 podStartE2EDuration="2.052938223s" podCreationTimestamp="2026-02-03 09:46:50 +0000 UTC" firstStartedPulling="2026-02-03 09:46:50.982010496 +0000 UTC m=+2202.132477871" lastFinishedPulling="2026-02-03 09:46:51.164969988 +0000 UTC m=+2202.315437373" observedRunningTime="2026-02-03 09:46:52.047699281 +0000 UTC m=+2203.198166666" watchObservedRunningTime="2026-02-03 09:46:52.052938223 +0000 UTC m=+2203.203405598" Feb 03 09:46:52 crc kubenswrapper[4756]: I0203 09:46:52.091172 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:54 crc kubenswrapper[4756]: I0203 09:46:54.613493 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:46:54 crc kubenswrapper[4756]: E0203 09:46:54.613989 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:46:54 crc kubenswrapper[4756]: I0203 09:46:54.810300 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rpp52"] Feb 03 09:46:54 crc kubenswrapper[4756]: I0203 09:46:54.810876 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rpp52" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerName="registry-server" containerID="cri-o://869021e66840c8836ab031f7fa86742657bcdd7bd29f35b425a1ee2cf95600ff" gracePeriod=2 Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.048543 4756 generic.go:334] "Generic (PLEG): container finished" podID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerID="869021e66840c8836ab031f7fa86742657bcdd7bd29f35b425a1ee2cf95600ff" exitCode=0 Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.048584 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp52" event={"ID":"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126","Type":"ContainerDied","Data":"869021e66840c8836ab031f7fa86742657bcdd7bd29f35b425a1ee2cf95600ff"} Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.308878 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.358215 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-utilities\") pod \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.358286 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx4wm\" (UniqueName: \"kubernetes.io/projected/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-kube-api-access-wx4wm\") pod \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.358401 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-catalog-content\") pod \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\" (UID: \"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126\") " Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.359343 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-utilities" (OuterVolumeSpecName: "utilities") pod "cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" (UID: "cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.364857 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-kube-api-access-wx4wm" (OuterVolumeSpecName: "kube-api-access-wx4wm") pod "cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" (UID: "cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126"). InnerVolumeSpecName "kube-api-access-wx4wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.410151 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" (UID: "cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.462760 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.462829 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx4wm\" (UniqueName: \"kubernetes.io/projected/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-kube-api-access-wx4wm\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:55 crc kubenswrapper[4756]: I0203 09:46:55.462847 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:46:56 crc kubenswrapper[4756]: I0203 09:46:56.057826 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp52" event={"ID":"cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126","Type":"ContainerDied","Data":"dc16bc5541c2cc7b8ceec0cc6f00ce24bf2e7f5caa12434d6c3bae3a0eabfc84"} Feb 03 09:46:56 crc kubenswrapper[4756]: I0203 09:46:56.057885 4756 scope.go:117] "RemoveContainer" containerID="869021e66840c8836ab031f7fa86742657bcdd7bd29f35b425a1ee2cf95600ff" Feb 03 09:46:56 crc kubenswrapper[4756]: I0203 09:46:56.057902 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpp52" Feb 03 09:46:56 crc kubenswrapper[4756]: I0203 09:46:56.076716 4756 scope.go:117] "RemoveContainer" containerID="731c05051c9ad49546659919b8e3dc3bb37e4e001915a500d9fecc2d8ed74dae" Feb 03 09:46:56 crc kubenswrapper[4756]: I0203 09:46:56.082379 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rpp52"] Feb 03 09:46:56 crc kubenswrapper[4756]: I0203 09:46:56.093547 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rpp52"] Feb 03 09:46:56 crc kubenswrapper[4756]: I0203 09:46:56.101422 4756 scope.go:117] "RemoveContainer" containerID="bc40ba2ba06cfc3d04dddc1acb214338ec886363973802a3ce35bfbba116e7b3" Feb 03 09:46:57 crc kubenswrapper[4756]: I0203 09:46:57.622627 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" path="/var/lib/kubelet/pods/cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126/volumes" Feb 03 09:47:08 crc kubenswrapper[4756]: I0203 09:47:08.614889 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:47:08 crc kubenswrapper[4756]: E0203 09:47:08.615745 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:47:19 crc kubenswrapper[4756]: I0203 09:47:19.619973 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:47:19 crc kubenswrapper[4756]: E0203 09:47:19.620812 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:47:32 crc kubenswrapper[4756]: I0203 09:47:32.614975 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:47:32 crc kubenswrapper[4756]: E0203 09:47:32.615783 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:47:45 crc kubenswrapper[4756]: I0203 09:47:45.614118 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:47:45 crc kubenswrapper[4756]: E0203 09:47:45.614961 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:47:59 crc kubenswrapper[4756]: I0203 09:47:59.635726 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:47:59 crc kubenswrapper[4756]: E0203 09:47:59.636685 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:48:12 crc kubenswrapper[4756]: I0203 09:48:12.615046 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:48:12 crc kubenswrapper[4756]: E0203 09:48:12.616051 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:48:27 crc kubenswrapper[4756]: I0203 09:48:27.615365 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:48:27 crc kubenswrapper[4756]: E0203 09:48:27.616772 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.823272 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mrgzw"] Feb 03 09:48:39 crc kubenswrapper[4756]: E0203 09:48:39.824225 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerName="registry-server" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.824239 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerName="registry-server" Feb 03 09:48:39 crc kubenswrapper[4756]: E0203 09:48:39.824250 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerName="extract-content" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.824256 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerName="extract-content" Feb 03 09:48:39 crc kubenswrapper[4756]: E0203 09:48:39.824279 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerName="extract-utilities" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.824286 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerName="extract-utilities" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.824473 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4fbe7f-21d3-4bc8-82a9-6e139b5f7126" containerName="registry-server" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.825859 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.838495 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mrgzw"] Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.941205 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28qxx\" (UniqueName: \"kubernetes.io/projected/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-kube-api-access-28qxx\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.941313 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-catalog-content\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:39 crc kubenswrapper[4756]: I0203 09:48:39.941978 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-utilities\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.043681 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28qxx\" (UniqueName: \"kubernetes.io/projected/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-kube-api-access-28qxx\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.043821 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-catalog-content\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.043847 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-utilities\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.044572 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-catalog-content\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.044636 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-utilities\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.067193 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28qxx\" (UniqueName: \"kubernetes.io/projected/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-kube-api-access-28qxx\") pod \"redhat-marketplace-mrgzw\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.162527 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.631380 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mrgzw"] Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.980055 4756 generic.go:334] "Generic (PLEG): container finished" podID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerID="770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba" exitCode=0 Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.980109 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mrgzw" event={"ID":"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b","Type":"ContainerDied","Data":"770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba"} Feb 03 09:48:40 crc kubenswrapper[4756]: I0203 09:48:40.980143 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mrgzw" event={"ID":"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b","Type":"ContainerStarted","Data":"50e92f097dabaae1c59e62ed6484aebb3c71b726f42c5db17ece1328d6af4abe"} Feb 03 09:48:41 crc kubenswrapper[4756]: I0203 09:48:41.614006 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:48:41 crc kubenswrapper[4756]: E0203 09:48:41.614613 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:48:41 crc kubenswrapper[4756]: I0203 09:48:41.988508 4756 generic.go:334] "Generic (PLEG): container finished" podID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerID="fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de" exitCode=0 Feb 03 09:48:41 crc kubenswrapper[4756]: I0203 09:48:41.988553 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mrgzw" event={"ID":"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b","Type":"ContainerDied","Data":"fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de"} Feb 03 09:48:43 crc kubenswrapper[4756]: I0203 09:48:43.000276 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mrgzw" event={"ID":"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b","Type":"ContainerStarted","Data":"5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94"} Feb 03 09:48:43 crc kubenswrapper[4756]: I0203 09:48:43.021511 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mrgzw" podStartSLOduration=2.597253695 podStartE2EDuration="4.021491263s" podCreationTimestamp="2026-02-03 09:48:39 +0000 UTC" firstStartedPulling="2026-02-03 09:48:40.985625791 +0000 UTC m=+2312.136093176" lastFinishedPulling="2026-02-03 09:48:42.409863369 +0000 UTC m=+2313.560330744" observedRunningTime="2026-02-03 09:48:43.019205942 +0000 UTC m=+2314.169673357" watchObservedRunningTime="2026-02-03 09:48:43.021491263 +0000 UTC m=+2314.171958648" Feb 03 09:48:50 crc kubenswrapper[4756]: I0203 09:48:50.062609 4756 generic.go:334] "Generic (PLEG): container finished" podID="0573a518-65e1-4717-a24e-285e30cad247" containerID="d52768fa1e2734e1c4d8d17e999a0ba1e0f7a38feae43da55a29ab607ed64dfa" exitCode=0 Feb 03 09:48:50 crc kubenswrapper[4756]: I0203 09:48:50.062880 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" event={"ID":"0573a518-65e1-4717-a24e-285e30cad247","Type":"ContainerDied","Data":"d52768fa1e2734e1c4d8d17e999a0ba1e0f7a38feae43da55a29ab607ed64dfa"} Feb 03 09:48:50 crc kubenswrapper[4756]: I0203 09:48:50.162682 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:50 crc kubenswrapper[4756]: I0203 09:48:50.162745 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:50 crc kubenswrapper[4756]: I0203 09:48:50.206051 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.119747 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.167916 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mrgzw"] Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.471264 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658102 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-ssh-key-openstack-edpm-ipam\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658229 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d72ck\" (UniqueName: \"kubernetes.io/projected/0573a518-65e1-4717-a24e-285e30cad247-kube-api-access-d72ck\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658315 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-inventory\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658428 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-0\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658511 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-0\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658568 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0573a518-65e1-4717-a24e-285e30cad247-nova-extra-config-0\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658645 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-1\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658687 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-combined-ca-bundle\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.658750 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-1\") pod \"0573a518-65e1-4717-a24e-285e30cad247\" (UID: \"0573a518-65e1-4717-a24e-285e30cad247\") " Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.663870 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0573a518-65e1-4717-a24e-285e30cad247-kube-api-access-d72ck" (OuterVolumeSpecName: "kube-api-access-d72ck") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "kube-api-access-d72ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.664092 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.682907 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0573a518-65e1-4717-a24e-285e30cad247-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.686841 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.689481 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.691374 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.692111 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.693337 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-inventory" (OuterVolumeSpecName: "inventory") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.696258 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "0573a518-65e1-4717-a24e-285e30cad247" (UID: "0573a518-65e1-4717-a24e-285e30cad247"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761769 4756 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761813 4756 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761829 4756 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/0573a518-65e1-4717-a24e-285e30cad247-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761843 4756 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761855 4756 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761868 4756 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761879 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761891 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d72ck\" (UniqueName: \"kubernetes.io/projected/0573a518-65e1-4717-a24e-285e30cad247-kube-api-access-d72ck\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:51 crc kubenswrapper[4756]: I0203 09:48:51.761903 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0573a518-65e1-4717-a24e-285e30cad247-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.080167 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.080179 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-k5dsp" event={"ID":"0573a518-65e1-4717-a24e-285e30cad247","Type":"ContainerDied","Data":"39a2447e07be04d4aed03fc6d3d096fb84772955dd47a596a40b4b5e34b3803f"} Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.080252 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39a2447e07be04d4aed03fc6d3d096fb84772955dd47a596a40b4b5e34b3803f" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.177403 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd"] Feb 03 09:48:52 crc kubenswrapper[4756]: E0203 09:48:52.177886 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0573a518-65e1-4717-a24e-285e30cad247" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.177905 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="0573a518-65e1-4717-a24e-285e30cad247" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.178116 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="0573a518-65e1-4717-a24e-285e30cad247" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.178761 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.181613 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.181951 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-f42hw" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.182257 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.182519 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.182778 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.186372 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd"] Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.372921 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.372985 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.373052 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.373073 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.373095 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.373121 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97j9b\" (UniqueName: \"kubernetes.io/projected/17726fdd-5344-437e-8060-5e26bcfb28cd-kube-api-access-97j9b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.373184 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.474921 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.475016 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.475039 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.475060 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.475081 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97j9b\" (UniqueName: \"kubernetes.io/projected/17726fdd-5344-437e-8060-5e26bcfb28cd-kube-api-access-97j9b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.475149 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.475195 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.478899 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.478899 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.483432 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.483688 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.484434 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.484753 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.521507 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97j9b\" (UniqueName: \"kubernetes.io/projected/17726fdd-5344-437e-8060-5e26bcfb28cd-kube-api-access-97j9b\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:52 crc kubenswrapper[4756]: I0203 09:48:52.820106 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.087968 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mrgzw" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerName="registry-server" containerID="cri-o://5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94" gracePeriod=2 Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.373595 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd"] Feb 03 09:48:53 crc kubenswrapper[4756]: W0203 09:48:53.378707 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17726fdd_5344_437e_8060_5e26bcfb28cd.slice/crio-9898324bf25b529ff2aa448c5fada990f6813389fbd4b1eedcf02620fac9a871 WatchSource:0}: Error finding container 9898324bf25b529ff2aa448c5fada990f6813389fbd4b1eedcf02620fac9a871: Status 404 returned error can't find the container with id 9898324bf25b529ff2aa448c5fada990f6813389fbd4b1eedcf02620fac9a871 Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.560584 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.694851 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28qxx\" (UniqueName: \"kubernetes.io/projected/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-kube-api-access-28qxx\") pod \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.695629 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-utilities\") pod \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.697101 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-utilities" (OuterVolumeSpecName: "utilities") pod "8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" (UID: "8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.697483 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-catalog-content\") pod \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\" (UID: \"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b\") " Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.698402 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.700938 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-kube-api-access-28qxx" (OuterVolumeSpecName: "kube-api-access-28qxx") pod "8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" (UID: "8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b"). InnerVolumeSpecName "kube-api-access-28qxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.719142 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" (UID: "8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.800886 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:53 crc kubenswrapper[4756]: I0203 09:48:53.800926 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28qxx\" (UniqueName: \"kubernetes.io/projected/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b-kube-api-access-28qxx\") on node \"crc\" DevicePath \"\"" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.098583 4756 generic.go:334] "Generic (PLEG): container finished" podID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerID="5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94" exitCode=0 Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.098692 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mrgzw" event={"ID":"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b","Type":"ContainerDied","Data":"5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94"} Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.098726 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mrgzw" event={"ID":"8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b","Type":"ContainerDied","Data":"50e92f097dabaae1c59e62ed6484aebb3c71b726f42c5db17ece1328d6af4abe"} Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.098760 4756 scope.go:117] "RemoveContainer" containerID="5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.098926 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mrgzw" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.105431 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" event={"ID":"17726fdd-5344-437e-8060-5e26bcfb28cd","Type":"ContainerStarted","Data":"a27d8610f4dab5a6a65ede888a00de99f9aae23bf5844806276449e4d26dc190"} Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.105500 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" event={"ID":"17726fdd-5344-437e-8060-5e26bcfb28cd","Type":"ContainerStarted","Data":"9898324bf25b529ff2aa448c5fada990f6813389fbd4b1eedcf02620fac9a871"} Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.126569 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" podStartSLOduration=1.926701311 podStartE2EDuration="2.126549907s" podCreationTimestamp="2026-02-03 09:48:52 +0000 UTC" firstStartedPulling="2026-02-03 09:48:53.38083581 +0000 UTC m=+2324.531303185" lastFinishedPulling="2026-02-03 09:48:53.580684406 +0000 UTC m=+2324.731151781" observedRunningTime="2026-02-03 09:48:54.123288346 +0000 UTC m=+2325.273755721" watchObservedRunningTime="2026-02-03 09:48:54.126549907 +0000 UTC m=+2325.277017282" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.126730 4756 scope.go:117] "RemoveContainer" containerID="fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.145703 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mrgzw"] Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.153898 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mrgzw"] Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.157860 4756 scope.go:117] "RemoveContainer" containerID="770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.187206 4756 scope.go:117] "RemoveContainer" containerID="5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94" Feb 03 09:48:54 crc kubenswrapper[4756]: E0203 09:48:54.187618 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94\": container with ID starting with 5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94 not found: ID does not exist" containerID="5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.187651 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94"} err="failed to get container status \"5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94\": rpc error: code = NotFound desc = could not find container \"5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94\": container with ID starting with 5e9f04d55776d6fc4235b34d8066f52bd1f6d8e4fdddc4621afacb6bd955bf94 not found: ID does not exist" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.187671 4756 scope.go:117] "RemoveContainer" containerID="fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de" Feb 03 09:48:54 crc kubenswrapper[4756]: E0203 09:48:54.187983 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de\": container with ID starting with fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de not found: ID does not exist" containerID="fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.188018 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de"} err="failed to get container status \"fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de\": rpc error: code = NotFound desc = could not find container \"fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de\": container with ID starting with fb1bdeb03f2bf5536214e909e3bbfad55fab33e2949d794ed5c2c4a04acac3de not found: ID does not exist" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.188037 4756 scope.go:117] "RemoveContainer" containerID="770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba" Feb 03 09:48:54 crc kubenswrapper[4756]: E0203 09:48:54.188284 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba\": container with ID starting with 770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba not found: ID does not exist" containerID="770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba" Feb 03 09:48:54 crc kubenswrapper[4756]: I0203 09:48:54.188354 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba"} err="failed to get container status \"770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba\": rpc error: code = NotFound desc = could not find container \"770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba\": container with ID starting with 770f26be264ae73487b2c13e866dc5d1f8501220cd01b242613b584cdc446cba not found: ID does not exist" Feb 03 09:48:55 crc kubenswrapper[4756]: I0203 09:48:55.614794 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:48:55 crc kubenswrapper[4756]: E0203 09:48:55.615276 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:48:55 crc kubenswrapper[4756]: I0203 09:48:55.626019 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" path="/var/lib/kubelet/pods/8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b/volumes" Feb 03 09:49:07 crc kubenswrapper[4756]: I0203 09:49:07.614378 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:49:07 crc kubenswrapper[4756]: E0203 09:49:07.615159 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:49:19 crc kubenswrapper[4756]: I0203 09:49:19.619150 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:49:19 crc kubenswrapper[4756]: E0203 09:49:19.620058 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:49:33 crc kubenswrapper[4756]: I0203 09:49:33.615225 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:49:33 crc kubenswrapper[4756]: E0203 09:49:33.616017 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:49:47 crc kubenswrapper[4756]: I0203 09:49:47.614930 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:49:47 crc kubenswrapper[4756]: E0203 09:49:47.615804 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:50:02 crc kubenswrapper[4756]: I0203 09:50:02.615054 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:50:02 crc kubenswrapper[4756]: E0203 09:50:02.616123 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:50:14 crc kubenswrapper[4756]: I0203 09:50:14.615507 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:50:14 crc kubenswrapper[4756]: E0203 09:50:14.616796 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:50:26 crc kubenswrapper[4756]: I0203 09:50:26.614249 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:50:26 crc kubenswrapper[4756]: E0203 09:50:26.615959 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:50:37 crc kubenswrapper[4756]: I0203 09:50:37.615005 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:50:37 crc kubenswrapper[4756]: E0203 09:50:37.616013 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:50:48 crc kubenswrapper[4756]: I0203 09:50:48.614870 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:50:48 crc kubenswrapper[4756]: E0203 09:50:48.617488 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:51:00 crc kubenswrapper[4756]: I0203 09:51:00.614683 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:51:00 crc kubenswrapper[4756]: E0203 09:51:00.615663 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:51:10 crc kubenswrapper[4756]: I0203 09:51:10.312804 4756 generic.go:334] "Generic (PLEG): container finished" podID="17726fdd-5344-437e-8060-5e26bcfb28cd" containerID="a27d8610f4dab5a6a65ede888a00de99f9aae23bf5844806276449e4d26dc190" exitCode=0 Feb 03 09:51:10 crc kubenswrapper[4756]: I0203 09:51:10.313071 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" event={"ID":"17726fdd-5344-437e-8060-5e26bcfb28cd","Type":"ContainerDied","Data":"a27d8610f4dab5a6a65ede888a00de99f9aae23bf5844806276449e4d26dc190"} Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.739826 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.852387 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-telemetry-combined-ca-bundle\") pod \"17726fdd-5344-437e-8060-5e26bcfb28cd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.852429 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-1\") pod \"17726fdd-5344-437e-8060-5e26bcfb28cd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.852621 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-2\") pod \"17726fdd-5344-437e-8060-5e26bcfb28cd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.852716 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-0\") pod \"17726fdd-5344-437e-8060-5e26bcfb28cd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.852768 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-inventory\") pod \"17726fdd-5344-437e-8060-5e26bcfb28cd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.852794 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97j9b\" (UniqueName: \"kubernetes.io/projected/17726fdd-5344-437e-8060-5e26bcfb28cd-kube-api-access-97j9b\") pod \"17726fdd-5344-437e-8060-5e26bcfb28cd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.852866 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ssh-key-openstack-edpm-ipam\") pod \"17726fdd-5344-437e-8060-5e26bcfb28cd\" (UID: \"17726fdd-5344-437e-8060-5e26bcfb28cd\") " Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.857774 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "17726fdd-5344-437e-8060-5e26bcfb28cd" (UID: "17726fdd-5344-437e-8060-5e26bcfb28cd"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.861626 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17726fdd-5344-437e-8060-5e26bcfb28cd-kube-api-access-97j9b" (OuterVolumeSpecName: "kube-api-access-97j9b") pod "17726fdd-5344-437e-8060-5e26bcfb28cd" (UID: "17726fdd-5344-437e-8060-5e26bcfb28cd"). InnerVolumeSpecName "kube-api-access-97j9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.883502 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-inventory" (OuterVolumeSpecName: "inventory") pod "17726fdd-5344-437e-8060-5e26bcfb28cd" (UID: "17726fdd-5344-437e-8060-5e26bcfb28cd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.883572 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "17726fdd-5344-437e-8060-5e26bcfb28cd" (UID: "17726fdd-5344-437e-8060-5e26bcfb28cd"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.887621 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "17726fdd-5344-437e-8060-5e26bcfb28cd" (UID: "17726fdd-5344-437e-8060-5e26bcfb28cd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.888398 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "17726fdd-5344-437e-8060-5e26bcfb28cd" (UID: "17726fdd-5344-437e-8060-5e26bcfb28cd"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.893324 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "17726fdd-5344-437e-8060-5e26bcfb28cd" (UID: "17726fdd-5344-437e-8060-5e26bcfb28cd"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.955097 4756 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.955149 4756 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-inventory\") on node \"crc\" DevicePath \"\"" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.955159 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97j9b\" (UniqueName: \"kubernetes.io/projected/17726fdd-5344-437e-8060-5e26bcfb28cd-kube-api-access-97j9b\") on node \"crc\" DevicePath \"\"" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.955170 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.955180 4756 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.955190 4756 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 03 09:51:11 crc kubenswrapper[4756]: I0203 09:51:11.955199 4756 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/17726fdd-5344-437e-8060-5e26bcfb28cd-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 03 09:51:12 crc kubenswrapper[4756]: I0203 09:51:12.353278 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" event={"ID":"17726fdd-5344-437e-8060-5e26bcfb28cd","Type":"ContainerDied","Data":"9898324bf25b529ff2aa448c5fada990f6813389fbd4b1eedcf02620fac9a871"} Feb 03 09:51:12 crc kubenswrapper[4756]: I0203 09:51:12.353326 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9898324bf25b529ff2aa448c5fada990f6813389fbd4b1eedcf02620fac9a871" Feb 03 09:51:12 crc kubenswrapper[4756]: I0203 09:51:12.353389 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd" Feb 03 09:51:12 crc kubenswrapper[4756]: I0203 09:51:12.615131 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:51:12 crc kubenswrapper[4756]: E0203 09:51:12.615657 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:51:26 crc kubenswrapper[4756]: I0203 09:51:26.614916 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:51:26 crc kubenswrapper[4756]: E0203 09:51:26.616175 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:51:37 crc kubenswrapper[4756]: I0203 09:51:37.614420 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:51:37 crc kubenswrapper[4756]: E0203 09:51:37.615251 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:51:52 crc kubenswrapper[4756]: I0203 09:51:52.614805 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:51:53 crc kubenswrapper[4756]: I0203 09:51:53.698782 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"1070612bc0b8f5b21089e24ef9c9fd6fc8929096b4acc377c5193d66c86497c1"} Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.403016 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 03 09:52:06 crc kubenswrapper[4756]: E0203 09:52:06.404226 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerName="registry-server" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.404248 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerName="registry-server" Feb 03 09:52:06 crc kubenswrapper[4756]: E0203 09:52:06.404265 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17726fdd-5344-437e-8060-5e26bcfb28cd" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.404277 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="17726fdd-5344-437e-8060-5e26bcfb28cd" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 03 09:52:06 crc kubenswrapper[4756]: E0203 09:52:06.404301 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerName="extract-content" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.404309 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerName="extract-content" Feb 03 09:52:06 crc kubenswrapper[4756]: E0203 09:52:06.404330 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerName="extract-utilities" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.404337 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerName="extract-utilities" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.404582 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="17726fdd-5344-437e-8060-5e26bcfb28cd" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.404602 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e2b7b5a-22f1-4738-8dfa-2bdbcb8b8a9b" containerName="registry-server" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.405421 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.408280 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.409309 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-m8rc4" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.410895 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.412953 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434610 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434660 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434689 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnhdq\" (UniqueName: \"kubernetes.io/projected/c529c965-e4c3-4e32-a5b2-445aebef0f08-kube-api-access-mnhdq\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434706 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434744 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434800 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-config-data\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434857 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434873 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.434925 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.436826 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536340 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536407 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536428 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536471 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536491 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnhdq\" (UniqueName: \"kubernetes.io/projected/c529c965-e4c3-4e32-a5b2-445aebef0f08-kube-api-access-mnhdq\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536530 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536587 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-config-data\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536645 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.536953 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.537035 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.537099 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.537408 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.538853 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-config-data\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.539985 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.545606 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.546305 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.552027 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.558606 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnhdq\" (UniqueName: \"kubernetes.io/projected/c529c965-e4c3-4e32-a5b2-445aebef0f08-kube-api-access-mnhdq\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.589623 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " pod="openstack/tempest-tests-tempest" Feb 03 09:52:06 crc kubenswrapper[4756]: I0203 09:52:06.735562 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 03 09:52:07 crc kubenswrapper[4756]: I0203 09:52:07.194936 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 03 09:52:07 crc kubenswrapper[4756]: I0203 09:52:07.198259 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 09:52:07 crc kubenswrapper[4756]: I0203 09:52:07.835923 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c529c965-e4c3-4e32-a5b2-445aebef0f08","Type":"ContainerStarted","Data":"2050a6e09bb8f32cfd7701c5bb271b4a0547f632780128ab2dc39b6786d070ca"} Feb 03 09:52:32 crc kubenswrapper[4756]: E0203 09:52:32.931008 4756 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 03 09:52:32 crc kubenswrapper[4756]: E0203 09:52:32.931925 4756 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mnhdq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(c529c965-e4c3-4e32-a5b2-445aebef0f08): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 03 09:52:32 crc kubenswrapper[4756]: E0203 09:52:32.933381 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="c529c965-e4c3-4e32-a5b2-445aebef0f08" Feb 03 09:52:33 crc kubenswrapper[4756]: E0203 09:52:33.109081 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="c529c965-e4c3-4e32-a5b2-445aebef0f08" Feb 03 09:52:45 crc kubenswrapper[4756]: I0203 09:52:45.091412 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 03 09:52:46 crc kubenswrapper[4756]: I0203 09:52:46.227485 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c529c965-e4c3-4e32-a5b2-445aebef0f08","Type":"ContainerStarted","Data":"9141af1444872117c0828db828b928b1f2dab3fc997730d8a50dff377eaecefe"} Feb 03 09:52:46 crc kubenswrapper[4756]: I0203 09:52:46.257030 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.365634603 podStartE2EDuration="41.257012286s" podCreationTimestamp="2026-02-03 09:52:05 +0000 UTC" firstStartedPulling="2026-02-03 09:52:07.198048557 +0000 UTC m=+2518.348515932" lastFinishedPulling="2026-02-03 09:52:45.08942624 +0000 UTC m=+2556.239893615" observedRunningTime="2026-02-03 09:52:46.253169575 +0000 UTC m=+2557.403636990" watchObservedRunningTime="2026-02-03 09:52:46.257012286 +0000 UTC m=+2557.407479661" Feb 03 09:54:13 crc kubenswrapper[4756]: I0203 09:54:13.566114 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:54:13 crc kubenswrapper[4756]: I0203 09:54:13.566740 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.050308 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bk8x4"] Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.057561 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.072555 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bk8x4"] Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.094180 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn8mc\" (UniqueName: \"kubernetes.io/projected/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-kube-api-access-kn8mc\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.094472 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-utilities\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.094534 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-catalog-content\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.196323 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn8mc\" (UniqueName: \"kubernetes.io/projected/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-kube-api-access-kn8mc\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.196688 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-utilities\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.196720 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-catalog-content\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.197281 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-catalog-content\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.197476 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-utilities\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.216326 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn8mc\" (UniqueName: \"kubernetes.io/projected/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-kube-api-access-kn8mc\") pod \"redhat-operators-bk8x4\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.406129 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:40 crc kubenswrapper[4756]: I0203 09:54:40.890005 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bk8x4"] Feb 03 09:54:40 crc kubenswrapper[4756]: W0203 09:54:40.902919 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a3be547_5fdc_4dcc_abc9_9b7410bbe692.slice/crio-2a90bb07cab326ab3c2b7cc0a6b1ae479d9af5d433f0739a4a33905f093b9a0b WatchSource:0}: Error finding container 2a90bb07cab326ab3c2b7cc0a6b1ae479d9af5d433f0739a4a33905f093b9a0b: Status 404 returned error can't find the container with id 2a90bb07cab326ab3c2b7cc0a6b1ae479d9af5d433f0739a4a33905f093b9a0b Feb 03 09:54:41 crc kubenswrapper[4756]: I0203 09:54:41.244161 4756 generic.go:334] "Generic (PLEG): container finished" podID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerID="3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081" exitCode=0 Feb 03 09:54:41 crc kubenswrapper[4756]: I0203 09:54:41.244220 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bk8x4" event={"ID":"5a3be547-5fdc-4dcc-abc9-9b7410bbe692","Type":"ContainerDied","Data":"3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081"} Feb 03 09:54:41 crc kubenswrapper[4756]: I0203 09:54:41.244559 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bk8x4" event={"ID":"5a3be547-5fdc-4dcc-abc9-9b7410bbe692","Type":"ContainerStarted","Data":"2a90bb07cab326ab3c2b7cc0a6b1ae479d9af5d433f0739a4a33905f093b9a0b"} Feb 03 09:54:42 crc kubenswrapper[4756]: I0203 09:54:42.256406 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bk8x4" event={"ID":"5a3be547-5fdc-4dcc-abc9-9b7410bbe692","Type":"ContainerStarted","Data":"3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834"} Feb 03 09:54:43 crc kubenswrapper[4756]: I0203 09:54:43.566172 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:54:43 crc kubenswrapper[4756]: I0203 09:54:43.566222 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:54:44 crc kubenswrapper[4756]: I0203 09:54:44.274951 4756 generic.go:334] "Generic (PLEG): container finished" podID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerID="3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834" exitCode=0 Feb 03 09:54:44 crc kubenswrapper[4756]: I0203 09:54:44.275051 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bk8x4" event={"ID":"5a3be547-5fdc-4dcc-abc9-9b7410bbe692","Type":"ContainerDied","Data":"3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834"} Feb 03 09:54:45 crc kubenswrapper[4756]: I0203 09:54:45.287936 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bk8x4" event={"ID":"5a3be547-5fdc-4dcc-abc9-9b7410bbe692","Type":"ContainerStarted","Data":"9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd"} Feb 03 09:54:45 crc kubenswrapper[4756]: I0203 09:54:45.313066 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bk8x4" podStartSLOduration=1.837965179 podStartE2EDuration="5.313050367s" podCreationTimestamp="2026-02-03 09:54:40 +0000 UTC" firstStartedPulling="2026-02-03 09:54:41.245875766 +0000 UTC m=+2672.396343151" lastFinishedPulling="2026-02-03 09:54:44.720960964 +0000 UTC m=+2675.871428339" observedRunningTime="2026-02-03 09:54:45.310130255 +0000 UTC m=+2676.460597630" watchObservedRunningTime="2026-02-03 09:54:45.313050367 +0000 UTC m=+2676.463517742" Feb 03 09:54:50 crc kubenswrapper[4756]: I0203 09:54:50.407886 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:50 crc kubenswrapper[4756]: I0203 09:54:50.408301 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:54:51 crc kubenswrapper[4756]: I0203 09:54:51.463001 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bk8x4" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="registry-server" probeResult="failure" output=< Feb 03 09:54:51 crc kubenswrapper[4756]: timeout: failed to connect service ":50051" within 1s Feb 03 09:54:51 crc kubenswrapper[4756]: > Feb 03 09:55:00 crc kubenswrapper[4756]: I0203 09:55:00.455957 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:55:00 crc kubenswrapper[4756]: I0203 09:55:00.505853 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:55:00 crc kubenswrapper[4756]: I0203 09:55:00.691351 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bk8x4"] Feb 03 09:55:02 crc kubenswrapper[4756]: I0203 09:55:02.437522 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bk8x4" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="registry-server" containerID="cri-o://9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd" gracePeriod=2 Feb 03 09:55:02 crc kubenswrapper[4756]: I0203 09:55:02.863565 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:55:02 crc kubenswrapper[4756]: I0203 09:55:02.939537 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-catalog-content\") pod \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " Feb 03 09:55:02 crc kubenswrapper[4756]: I0203 09:55:02.940009 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-utilities\") pod \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " Feb 03 09:55:02 crc kubenswrapper[4756]: I0203 09:55:02.940082 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kn8mc\" (UniqueName: \"kubernetes.io/projected/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-kube-api-access-kn8mc\") pod \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\" (UID: \"5a3be547-5fdc-4dcc-abc9-9b7410bbe692\") " Feb 03 09:55:02 crc kubenswrapper[4756]: I0203 09:55:02.940752 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-utilities" (OuterVolumeSpecName: "utilities") pod "5a3be547-5fdc-4dcc-abc9-9b7410bbe692" (UID: "5a3be547-5fdc-4dcc-abc9-9b7410bbe692"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:55:02 crc kubenswrapper[4756]: I0203 09:55:02.948836 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-kube-api-access-kn8mc" (OuterVolumeSpecName: "kube-api-access-kn8mc") pod "5a3be547-5fdc-4dcc-abc9-9b7410bbe692" (UID: "5a3be547-5fdc-4dcc-abc9-9b7410bbe692"). InnerVolumeSpecName "kube-api-access-kn8mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.042560 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.042590 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kn8mc\" (UniqueName: \"kubernetes.io/projected/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-kube-api-access-kn8mc\") on node \"crc\" DevicePath \"\"" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.054839 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a3be547-5fdc-4dcc-abc9-9b7410bbe692" (UID: "5a3be547-5fdc-4dcc-abc9-9b7410bbe692"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.145065 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a3be547-5fdc-4dcc-abc9-9b7410bbe692-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.449214 4756 generic.go:334] "Generic (PLEG): container finished" podID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerID="9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd" exitCode=0 Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.449271 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bk8x4" event={"ID":"5a3be547-5fdc-4dcc-abc9-9b7410bbe692","Type":"ContainerDied","Data":"9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd"} Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.449342 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bk8x4" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.450022 4756 scope.go:117] "RemoveContainer" containerID="9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.450003 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bk8x4" event={"ID":"5a3be547-5fdc-4dcc-abc9-9b7410bbe692","Type":"ContainerDied","Data":"2a90bb07cab326ab3c2b7cc0a6b1ae479d9af5d433f0739a4a33905f093b9a0b"} Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.471103 4756 scope.go:117] "RemoveContainer" containerID="3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.499067 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bk8x4"] Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.509110 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bk8x4"] Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.527157 4756 scope.go:117] "RemoveContainer" containerID="3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.575692 4756 scope.go:117] "RemoveContainer" containerID="9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd" Feb 03 09:55:03 crc kubenswrapper[4756]: E0203 09:55:03.576245 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd\": container with ID starting with 9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd not found: ID does not exist" containerID="9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.576363 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd"} err="failed to get container status \"9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd\": rpc error: code = NotFound desc = could not find container \"9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd\": container with ID starting with 9e94ce331bce88f8f4d13c4a6a88ce5a1e9c8017620fd5f7525845ae02cfd4bd not found: ID does not exist" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.576471 4756 scope.go:117] "RemoveContainer" containerID="3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834" Feb 03 09:55:03 crc kubenswrapper[4756]: E0203 09:55:03.576933 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834\": container with ID starting with 3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834 not found: ID does not exist" containerID="3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.577027 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834"} err="failed to get container status \"3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834\": rpc error: code = NotFound desc = could not find container \"3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834\": container with ID starting with 3eeacb2d844f32a8f11d07d7a207ba76044cb347f41fa1b8b94410ccce5df834 not found: ID does not exist" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.577095 4756 scope.go:117] "RemoveContainer" containerID="3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081" Feb 03 09:55:03 crc kubenswrapper[4756]: E0203 09:55:03.577558 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081\": container with ID starting with 3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081 not found: ID does not exist" containerID="3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.577609 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081"} err="failed to get container status \"3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081\": rpc error: code = NotFound desc = could not find container \"3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081\": container with ID starting with 3e29450f86b7c1d1e705a15695848924a891f49ca0b1d3c5a340b3376a4ce081 not found: ID does not exist" Feb 03 09:55:03 crc kubenswrapper[4756]: I0203 09:55:03.624750 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" path="/var/lib/kubelet/pods/5a3be547-5fdc-4dcc-abc9-9b7410bbe692/volumes" Feb 03 09:55:13 crc kubenswrapper[4756]: I0203 09:55:13.570100 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:55:13 crc kubenswrapper[4756]: I0203 09:55:13.570784 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:55:13 crc kubenswrapper[4756]: I0203 09:55:13.570840 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:55:13 crc kubenswrapper[4756]: I0203 09:55:13.571662 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1070612bc0b8f5b21089e24ef9c9fd6fc8929096b4acc377c5193d66c86497c1"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:55:13 crc kubenswrapper[4756]: I0203 09:55:13.571729 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://1070612bc0b8f5b21089e24ef9c9fd6fc8929096b4acc377c5193d66c86497c1" gracePeriod=600 Feb 03 09:55:14 crc kubenswrapper[4756]: I0203 09:55:14.559665 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="1070612bc0b8f5b21089e24ef9c9fd6fc8929096b4acc377c5193d66c86497c1" exitCode=0 Feb 03 09:55:14 crc kubenswrapper[4756]: I0203 09:55:14.559716 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"1070612bc0b8f5b21089e24ef9c9fd6fc8929096b4acc377c5193d66c86497c1"} Feb 03 09:55:14 crc kubenswrapper[4756]: I0203 09:55:14.560032 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b"} Feb 03 09:55:14 crc kubenswrapper[4756]: I0203 09:55:14.560052 4756 scope.go:117] "RemoveContainer" containerID="3c329d2fe97cf90575f45f2d5fc24cfcc0dcfe8a05edc75cec8846fde8b60945" Feb 03 09:57:13 crc kubenswrapper[4756]: I0203 09:57:13.565892 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:57:13 crc kubenswrapper[4756]: I0203 09:57:13.566484 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.104945 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ws2vb"] Feb 03 09:57:25 crc kubenswrapper[4756]: E0203 09:57:25.106030 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="extract-content" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.106048 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="extract-content" Feb 03 09:57:25 crc kubenswrapper[4756]: E0203 09:57:25.106073 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="registry-server" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.106083 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="registry-server" Feb 03 09:57:25 crc kubenswrapper[4756]: E0203 09:57:25.106106 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="extract-utilities" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.106115 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="extract-utilities" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.106368 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a3be547-5fdc-4dcc-abc9-9b7410bbe692" containerName="registry-server" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.108533 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.135099 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ws2vb"] Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.255989 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8p5g\" (UniqueName: \"kubernetes.io/projected/f969762e-ae35-4f20-9efe-7656e8a2d9ff-kube-api-access-f8p5g\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.256422 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-catalog-content\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.256499 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-utilities\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.358542 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-catalog-content\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.358622 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-utilities\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.358722 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8p5g\" (UniqueName: \"kubernetes.io/projected/f969762e-ae35-4f20-9efe-7656e8a2d9ff-kube-api-access-f8p5g\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.359205 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-catalog-content\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.359218 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-utilities\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.393984 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8p5g\" (UniqueName: \"kubernetes.io/projected/f969762e-ae35-4f20-9efe-7656e8a2d9ff-kube-api-access-f8p5g\") pod \"community-operators-ws2vb\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:25 crc kubenswrapper[4756]: I0203 09:57:25.430396 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:26 crc kubenswrapper[4756]: I0203 09:57:26.005141 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ws2vb"] Feb 03 09:57:26 crc kubenswrapper[4756]: I0203 09:57:26.681639 4756 generic.go:334] "Generic (PLEG): container finished" podID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerID="3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e" exitCode=0 Feb 03 09:57:26 crc kubenswrapper[4756]: I0203 09:57:26.681925 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws2vb" event={"ID":"f969762e-ae35-4f20-9efe-7656e8a2d9ff","Type":"ContainerDied","Data":"3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e"} Feb 03 09:57:26 crc kubenswrapper[4756]: I0203 09:57:26.682089 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws2vb" event={"ID":"f969762e-ae35-4f20-9efe-7656e8a2d9ff","Type":"ContainerStarted","Data":"1904433e41bc89f5e9e177735b1e91093c9b8a54d5dfe35803e368f257d845a5"} Feb 03 09:57:26 crc kubenswrapper[4756]: I0203 09:57:26.683981 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 09:57:27 crc kubenswrapper[4756]: I0203 09:57:27.692615 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws2vb" event={"ID":"f969762e-ae35-4f20-9efe-7656e8a2d9ff","Type":"ContainerStarted","Data":"753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e"} Feb 03 09:57:29 crc kubenswrapper[4756]: I0203 09:57:29.717113 4756 generic.go:334] "Generic (PLEG): container finished" podID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerID="753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e" exitCode=0 Feb 03 09:57:29 crc kubenswrapper[4756]: I0203 09:57:29.717181 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws2vb" event={"ID":"f969762e-ae35-4f20-9efe-7656e8a2d9ff","Type":"ContainerDied","Data":"753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e"} Feb 03 09:57:30 crc kubenswrapper[4756]: I0203 09:57:30.729696 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws2vb" event={"ID":"f969762e-ae35-4f20-9efe-7656e8a2d9ff","Type":"ContainerStarted","Data":"646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded"} Feb 03 09:57:30 crc kubenswrapper[4756]: I0203 09:57:30.752318 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ws2vb" podStartSLOduration=2.156005794 podStartE2EDuration="5.752297966s" podCreationTimestamp="2026-02-03 09:57:25 +0000 UTC" firstStartedPulling="2026-02-03 09:57:26.683636318 +0000 UTC m=+2837.834103703" lastFinishedPulling="2026-02-03 09:57:30.2799285 +0000 UTC m=+2841.430395875" observedRunningTime="2026-02-03 09:57:30.750168669 +0000 UTC m=+2841.900636044" watchObservedRunningTime="2026-02-03 09:57:30.752297966 +0000 UTC m=+2841.902765351" Feb 03 09:57:35 crc kubenswrapper[4756]: I0203 09:57:35.431287 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:35 crc kubenswrapper[4756]: I0203 09:57:35.432056 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:35 crc kubenswrapper[4756]: I0203 09:57:35.494422 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:35 crc kubenswrapper[4756]: I0203 09:57:35.831475 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:35 crc kubenswrapper[4756]: I0203 09:57:35.889209 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ws2vb"] Feb 03 09:57:37 crc kubenswrapper[4756]: I0203 09:57:37.815062 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ws2vb" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerName="registry-server" containerID="cri-o://646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded" gracePeriod=2 Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.341562 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.438006 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8p5g\" (UniqueName: \"kubernetes.io/projected/f969762e-ae35-4f20-9efe-7656e8a2d9ff-kube-api-access-f8p5g\") pod \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.438215 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-catalog-content\") pod \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.438421 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-utilities\") pod \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\" (UID: \"f969762e-ae35-4f20-9efe-7656e8a2d9ff\") " Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.439172 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-utilities" (OuterVolumeSpecName: "utilities") pod "f969762e-ae35-4f20-9efe-7656e8a2d9ff" (UID: "f969762e-ae35-4f20-9efe-7656e8a2d9ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.444690 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f969762e-ae35-4f20-9efe-7656e8a2d9ff-kube-api-access-f8p5g" (OuterVolumeSpecName: "kube-api-access-f8p5g") pod "f969762e-ae35-4f20-9efe-7656e8a2d9ff" (UID: "f969762e-ae35-4f20-9efe-7656e8a2d9ff"). InnerVolumeSpecName "kube-api-access-f8p5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.490275 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f969762e-ae35-4f20-9efe-7656e8a2d9ff" (UID: "f969762e-ae35-4f20-9efe-7656e8a2d9ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.540367 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.540405 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8p5g\" (UniqueName: \"kubernetes.io/projected/f969762e-ae35-4f20-9efe-7656e8a2d9ff-kube-api-access-f8p5g\") on node \"crc\" DevicePath \"\"" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.540419 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f969762e-ae35-4f20-9efe-7656e8a2d9ff-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.830743 4756 generic.go:334] "Generic (PLEG): container finished" podID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerID="646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded" exitCode=0 Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.830817 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ws2vb" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.830848 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws2vb" event={"ID":"f969762e-ae35-4f20-9efe-7656e8a2d9ff","Type":"ContainerDied","Data":"646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded"} Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.832996 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ws2vb" event={"ID":"f969762e-ae35-4f20-9efe-7656e8a2d9ff","Type":"ContainerDied","Data":"1904433e41bc89f5e9e177735b1e91093c9b8a54d5dfe35803e368f257d845a5"} Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.833052 4756 scope.go:117] "RemoveContainer" containerID="646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.856989 4756 scope.go:117] "RemoveContainer" containerID="753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.883800 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ws2vb"] Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.894132 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ws2vb"] Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.905974 4756 scope.go:117] "RemoveContainer" containerID="3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.934022 4756 scope.go:117] "RemoveContainer" containerID="646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded" Feb 03 09:57:38 crc kubenswrapper[4756]: E0203 09:57:38.935066 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded\": container with ID starting with 646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded not found: ID does not exist" containerID="646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.935104 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded"} err="failed to get container status \"646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded\": rpc error: code = NotFound desc = could not find container \"646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded\": container with ID starting with 646eb1478d3474fd42ddd3e560ddc8a46931a6a35f9da70b1ea4a2123a6e8ded not found: ID does not exist" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.935133 4756 scope.go:117] "RemoveContainer" containerID="753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e" Feb 03 09:57:38 crc kubenswrapper[4756]: E0203 09:57:38.935399 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e\": container with ID starting with 753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e not found: ID does not exist" containerID="753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.935426 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e"} err="failed to get container status \"753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e\": rpc error: code = NotFound desc = could not find container \"753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e\": container with ID starting with 753d54818e74bfa9a532ac5749696e5bfb70d1a07d95404cf36f56541b6c118e not found: ID does not exist" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.935476 4756 scope.go:117] "RemoveContainer" containerID="3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e" Feb 03 09:57:38 crc kubenswrapper[4756]: E0203 09:57:38.935827 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e\": container with ID starting with 3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e not found: ID does not exist" containerID="3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e" Feb 03 09:57:38 crc kubenswrapper[4756]: I0203 09:57:38.935856 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e"} err="failed to get container status \"3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e\": rpc error: code = NotFound desc = could not find container \"3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e\": container with ID starting with 3d7c4deaef11219668954acf91968e43dd6ddb9ec5ec9e9066fe753c56f64a8e not found: ID does not exist" Feb 03 09:57:39 crc kubenswrapper[4756]: I0203 09:57:39.625093 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" path="/var/lib/kubelet/pods/f969762e-ae35-4f20-9efe-7656e8a2d9ff/volumes" Feb 03 09:57:43 crc kubenswrapper[4756]: I0203 09:57:43.566802 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:57:43 crc kubenswrapper[4756]: I0203 09:57:43.568177 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:58:13 crc kubenswrapper[4756]: I0203 09:58:13.566056 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 09:58:13 crc kubenswrapper[4756]: I0203 09:58:13.567577 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 09:58:13 crc kubenswrapper[4756]: I0203 09:58:13.567985 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 09:58:13 crc kubenswrapper[4756]: I0203 09:58:13.568884 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 09:58:13 crc kubenswrapper[4756]: I0203 09:58:13.568958 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" gracePeriod=600 Feb 03 09:58:13 crc kubenswrapper[4756]: E0203 09:58:13.694816 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:58:14 crc kubenswrapper[4756]: I0203 09:58:14.160482 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" exitCode=0 Feb 03 09:58:14 crc kubenswrapper[4756]: I0203 09:58:14.160538 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b"} Feb 03 09:58:14 crc kubenswrapper[4756]: I0203 09:58:14.160630 4756 scope.go:117] "RemoveContainer" containerID="1070612bc0b8f5b21089e24ef9c9fd6fc8929096b4acc377c5193d66c86497c1" Feb 03 09:58:14 crc kubenswrapper[4756]: I0203 09:58:14.161385 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:58:14 crc kubenswrapper[4756]: E0203 09:58:14.161674 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:58:25 crc kubenswrapper[4756]: I0203 09:58:25.615311 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:58:25 crc kubenswrapper[4756]: E0203 09:58:25.616230 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:58:38 crc kubenswrapper[4756]: I0203 09:58:38.614105 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:58:38 crc kubenswrapper[4756]: E0203 09:58:38.615054 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:58:49 crc kubenswrapper[4756]: I0203 09:58:49.624235 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:58:49 crc kubenswrapper[4756]: E0203 09:58:49.625207 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:59:03 crc kubenswrapper[4756]: I0203 09:59:03.614895 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:59:03 crc kubenswrapper[4756]: E0203 09:59:03.615700 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:59:18 crc kubenswrapper[4756]: I0203 09:59:18.614961 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:59:18 crc kubenswrapper[4756]: E0203 09:59:18.616604 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:59:31 crc kubenswrapper[4756]: I0203 09:59:31.616013 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:59:31 crc kubenswrapper[4756]: E0203 09:59:31.616820 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:59:46 crc kubenswrapper[4756]: I0203 09:59:46.614285 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:59:46 crc kubenswrapper[4756]: E0203 09:59:46.615230 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.437476 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-442lm"] Feb 03 09:59:53 crc kubenswrapper[4756]: E0203 09:59:53.438960 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerName="extract-content" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.438984 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerName="extract-content" Feb 03 09:59:53 crc kubenswrapper[4756]: E0203 09:59:53.438998 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerName="registry-server" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.439009 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerName="registry-server" Feb 03 09:59:53 crc kubenswrapper[4756]: E0203 09:59:53.439026 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerName="extract-utilities" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.439039 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerName="extract-utilities" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.439690 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f969762e-ae35-4f20-9efe-7656e8a2d9ff" containerName="registry-server" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.441890 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.460235 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-442lm"] Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.588983 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-utilities\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.589686 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-catalog-content\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.589736 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htdnd\" (UniqueName: \"kubernetes.io/projected/7c3429c5-a971-48ef-a420-94f6f1717db8-kube-api-access-htdnd\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.692278 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-catalog-content\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.692356 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htdnd\" (UniqueName: \"kubernetes.io/projected/7c3429c5-a971-48ef-a420-94f6f1717db8-kube-api-access-htdnd\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.692484 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-utilities\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.692876 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-catalog-content\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.693325 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-utilities\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.713147 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htdnd\" (UniqueName: \"kubernetes.io/projected/7c3429c5-a971-48ef-a420-94f6f1717db8-kube-api-access-htdnd\") pod \"redhat-marketplace-442lm\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:53 crc kubenswrapper[4756]: I0203 09:59:53.765577 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 09:59:54 crc kubenswrapper[4756]: I0203 09:59:54.243488 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-442lm"] Feb 03 09:59:55 crc kubenswrapper[4756]: I0203 09:59:55.009647 4756 generic.go:334] "Generic (PLEG): container finished" podID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerID="334b863a20c3382d853a029fe5153d46cd9dc01c35d18fb692ff0d92213c12f1" exitCode=0 Feb 03 09:59:55 crc kubenswrapper[4756]: I0203 09:59:55.009758 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-442lm" event={"ID":"7c3429c5-a971-48ef-a420-94f6f1717db8","Type":"ContainerDied","Data":"334b863a20c3382d853a029fe5153d46cd9dc01c35d18fb692ff0d92213c12f1"} Feb 03 09:59:55 crc kubenswrapper[4756]: I0203 09:59:55.010274 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-442lm" event={"ID":"7c3429c5-a971-48ef-a420-94f6f1717db8","Type":"ContainerStarted","Data":"66dd6bdf39e51685a5d68a1d3e0cdbf060e8ceed8e6ed9e1eb787367bcba1055"} Feb 03 09:59:56 crc kubenswrapper[4756]: I0203 09:59:56.021356 4756 generic.go:334] "Generic (PLEG): container finished" podID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerID="2a44633565104b5b007a3120f840a2d54b716331726e0f7d9818b9d52593438e" exitCode=0 Feb 03 09:59:56 crc kubenswrapper[4756]: I0203 09:59:56.021426 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-442lm" event={"ID":"7c3429c5-a971-48ef-a420-94f6f1717db8","Type":"ContainerDied","Data":"2a44633565104b5b007a3120f840a2d54b716331726e0f7d9818b9d52593438e"} Feb 03 09:59:57 crc kubenswrapper[4756]: I0203 09:59:57.032669 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-442lm" event={"ID":"7c3429c5-a971-48ef-a420-94f6f1717db8","Type":"ContainerStarted","Data":"fdd7b8a85a2a3116c0b476fe64252899fee5dd574ceebd3af88e782c7f70d598"} Feb 03 09:59:57 crc kubenswrapper[4756]: I0203 09:59:57.058470 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-442lm" podStartSLOduration=2.570774653 podStartE2EDuration="4.058432649s" podCreationTimestamp="2026-02-03 09:59:53 +0000 UTC" firstStartedPulling="2026-02-03 09:59:55.012104992 +0000 UTC m=+2986.162572367" lastFinishedPulling="2026-02-03 09:59:56.499762988 +0000 UTC m=+2987.650230363" observedRunningTime="2026-02-03 09:59:57.050893373 +0000 UTC m=+2988.201360768" watchObservedRunningTime="2026-02-03 09:59:57.058432649 +0000 UTC m=+2988.208900024" Feb 03 09:59:59 crc kubenswrapper[4756]: I0203 09:59:59.620518 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 09:59:59 crc kubenswrapper[4756]: E0203 09:59:59.622142 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.146085 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr"] Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.148675 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.150809 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.151731 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.158174 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr"] Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.314050 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjjff\" (UniqueName: \"kubernetes.io/projected/81294c99-0854-4af1-b552-41f9695aa7bc-kube-api-access-cjjff\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.314591 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81294c99-0854-4af1-b552-41f9695aa7bc-config-volume\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.314640 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81294c99-0854-4af1-b552-41f9695aa7bc-secret-volume\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.417318 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81294c99-0854-4af1-b552-41f9695aa7bc-config-volume\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.417810 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81294c99-0854-4af1-b552-41f9695aa7bc-secret-volume\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.418002 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjjff\" (UniqueName: \"kubernetes.io/projected/81294c99-0854-4af1-b552-41f9695aa7bc-kube-api-access-cjjff\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.418414 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81294c99-0854-4af1-b552-41f9695aa7bc-config-volume\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.426308 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81294c99-0854-4af1-b552-41f9695aa7bc-secret-volume\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.435200 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjjff\" (UniqueName: \"kubernetes.io/projected/81294c99-0854-4af1-b552-41f9695aa7bc-kube-api-access-cjjff\") pod \"collect-profiles-29501880-q75lr\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.470102 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:00 crc kubenswrapper[4756]: I0203 10:00:00.943800 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr"] Feb 03 10:00:00 crc kubenswrapper[4756]: W0203 10:00:00.947172 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81294c99_0854_4af1_b552_41f9695aa7bc.slice/crio-0b51bb609393522971b7b9c39a8f3675764eb5d6c0883cdc3abda2fe67eb5502 WatchSource:0}: Error finding container 0b51bb609393522971b7b9c39a8f3675764eb5d6c0883cdc3abda2fe67eb5502: Status 404 returned error can't find the container with id 0b51bb609393522971b7b9c39a8f3675764eb5d6c0883cdc3abda2fe67eb5502 Feb 03 10:00:01 crc kubenswrapper[4756]: I0203 10:00:01.065416 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" event={"ID":"81294c99-0854-4af1-b552-41f9695aa7bc","Type":"ContainerStarted","Data":"0b51bb609393522971b7b9c39a8f3675764eb5d6c0883cdc3abda2fe67eb5502"} Feb 03 10:00:02 crc kubenswrapper[4756]: I0203 10:00:02.077035 4756 generic.go:334] "Generic (PLEG): container finished" podID="81294c99-0854-4af1-b552-41f9695aa7bc" containerID="c51c4a8e6d96ef6bfeb1404240e31aa9c4d6cf38e9d60bc1b292c3929975cb39" exitCode=0 Feb 03 10:00:02 crc kubenswrapper[4756]: I0203 10:00:02.077109 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" event={"ID":"81294c99-0854-4af1-b552-41f9695aa7bc","Type":"ContainerDied","Data":"c51c4a8e6d96ef6bfeb1404240e31aa9c4d6cf38e9d60bc1b292c3929975cb39"} Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.521376 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.685364 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81294c99-0854-4af1-b552-41f9695aa7bc-secret-volume\") pod \"81294c99-0854-4af1-b552-41f9695aa7bc\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.685733 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81294c99-0854-4af1-b552-41f9695aa7bc-config-volume\") pod \"81294c99-0854-4af1-b552-41f9695aa7bc\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.685819 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjjff\" (UniqueName: \"kubernetes.io/projected/81294c99-0854-4af1-b552-41f9695aa7bc-kube-api-access-cjjff\") pod \"81294c99-0854-4af1-b552-41f9695aa7bc\" (UID: \"81294c99-0854-4af1-b552-41f9695aa7bc\") " Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.686187 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81294c99-0854-4af1-b552-41f9695aa7bc-config-volume" (OuterVolumeSpecName: "config-volume") pod "81294c99-0854-4af1-b552-41f9695aa7bc" (UID: "81294c99-0854-4af1-b552-41f9695aa7bc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.691519 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81294c99-0854-4af1-b552-41f9695aa7bc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "81294c99-0854-4af1-b552-41f9695aa7bc" (UID: "81294c99-0854-4af1-b552-41f9695aa7bc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.692100 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81294c99-0854-4af1-b552-41f9695aa7bc-kube-api-access-cjjff" (OuterVolumeSpecName: "kube-api-access-cjjff") pod "81294c99-0854-4af1-b552-41f9695aa7bc" (UID: "81294c99-0854-4af1-b552-41f9695aa7bc"). InnerVolumeSpecName "kube-api-access-cjjff". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.766600 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.766671 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.788664 4756 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81294c99-0854-4af1-b552-41f9695aa7bc-config-volume\") on node \"crc\" DevicePath \"\"" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.788949 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjjff\" (UniqueName: \"kubernetes.io/projected/81294c99-0854-4af1-b552-41f9695aa7bc-kube-api-access-cjjff\") on node \"crc\" DevicePath \"\"" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.788977 4756 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81294c99-0854-4af1-b552-41f9695aa7bc-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 03 10:00:03 crc kubenswrapper[4756]: I0203 10:00:03.825269 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 10:00:04 crc kubenswrapper[4756]: I0203 10:00:04.095786 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" event={"ID":"81294c99-0854-4af1-b552-41f9695aa7bc","Type":"ContainerDied","Data":"0b51bb609393522971b7b9c39a8f3675764eb5d6c0883cdc3abda2fe67eb5502"} Feb 03 10:00:04 crc kubenswrapper[4756]: I0203 10:00:04.095834 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501880-q75lr" Feb 03 10:00:04 crc kubenswrapper[4756]: I0203 10:00:04.095849 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b51bb609393522971b7b9c39a8f3675764eb5d6c0883cdc3abda2fe67eb5502" Feb 03 10:00:04 crc kubenswrapper[4756]: I0203 10:00:04.150255 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 10:00:04 crc kubenswrapper[4756]: I0203 10:00:04.203911 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-442lm"] Feb 03 10:00:04 crc kubenswrapper[4756]: I0203 10:00:04.597094 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn"] Feb 03 10:00:04 crc kubenswrapper[4756]: I0203 10:00:04.607173 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501835-p9bwn"] Feb 03 10:00:05 crc kubenswrapper[4756]: I0203 10:00:05.652927 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac16e184-042c-43c9-8848-b549fb5e44a2" path="/var/lib/kubelet/pods/ac16e184-042c-43c9-8848-b549fb5e44a2/volumes" Feb 03 10:00:06 crc kubenswrapper[4756]: I0203 10:00:06.114822 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-442lm" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerName="registry-server" containerID="cri-o://fdd7b8a85a2a3116c0b476fe64252899fee5dd574ceebd3af88e782c7f70d598" gracePeriod=2 Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.127303 4756 generic.go:334] "Generic (PLEG): container finished" podID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerID="fdd7b8a85a2a3116c0b476fe64252899fee5dd574ceebd3af88e782c7f70d598" exitCode=0 Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.127491 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-442lm" event={"ID":"7c3429c5-a971-48ef-a420-94f6f1717db8","Type":"ContainerDied","Data":"fdd7b8a85a2a3116c0b476fe64252899fee5dd574ceebd3af88e782c7f70d598"} Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.399514 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.571861 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-catalog-content\") pod \"7c3429c5-a971-48ef-a420-94f6f1717db8\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.572171 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-utilities\") pod \"7c3429c5-a971-48ef-a420-94f6f1717db8\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.572228 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htdnd\" (UniqueName: \"kubernetes.io/projected/7c3429c5-a971-48ef-a420-94f6f1717db8-kube-api-access-htdnd\") pod \"7c3429c5-a971-48ef-a420-94f6f1717db8\" (UID: \"7c3429c5-a971-48ef-a420-94f6f1717db8\") " Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.573099 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-utilities" (OuterVolumeSpecName: "utilities") pod "7c3429c5-a971-48ef-a420-94f6f1717db8" (UID: "7c3429c5-a971-48ef-a420-94f6f1717db8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.573925 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.581284 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3429c5-a971-48ef-a420-94f6f1717db8-kube-api-access-htdnd" (OuterVolumeSpecName: "kube-api-access-htdnd") pod "7c3429c5-a971-48ef-a420-94f6f1717db8" (UID: "7c3429c5-a971-48ef-a420-94f6f1717db8"). InnerVolumeSpecName "kube-api-access-htdnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.599389 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c3429c5-a971-48ef-a420-94f6f1717db8" (UID: "7c3429c5-a971-48ef-a420-94f6f1717db8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.676229 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htdnd\" (UniqueName: \"kubernetes.io/projected/7c3429c5-a971-48ef-a420-94f6f1717db8-kube-api-access-htdnd\") on node \"crc\" DevicePath \"\"" Feb 03 10:00:07 crc kubenswrapper[4756]: I0203 10:00:07.676277 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3429c5-a971-48ef-a420-94f6f1717db8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 10:00:08 crc kubenswrapper[4756]: I0203 10:00:08.143959 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-442lm" event={"ID":"7c3429c5-a971-48ef-a420-94f6f1717db8","Type":"ContainerDied","Data":"66dd6bdf39e51685a5d68a1d3e0cdbf060e8ceed8e6ed9e1eb787367bcba1055"} Feb 03 10:00:08 crc kubenswrapper[4756]: I0203 10:00:08.144031 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-442lm" Feb 03 10:00:08 crc kubenswrapper[4756]: I0203 10:00:08.144330 4756 scope.go:117] "RemoveContainer" containerID="fdd7b8a85a2a3116c0b476fe64252899fee5dd574ceebd3af88e782c7f70d598" Feb 03 10:00:08 crc kubenswrapper[4756]: I0203 10:00:08.182814 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-442lm"] Feb 03 10:00:08 crc kubenswrapper[4756]: I0203 10:00:08.192541 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-442lm"] Feb 03 10:00:08 crc kubenswrapper[4756]: I0203 10:00:08.193619 4756 scope.go:117] "RemoveContainer" containerID="2a44633565104b5b007a3120f840a2d54b716331726e0f7d9818b9d52593438e" Feb 03 10:00:08 crc kubenswrapper[4756]: I0203 10:00:08.216692 4756 scope.go:117] "RemoveContainer" containerID="334b863a20c3382d853a029fe5153d46cd9dc01c35d18fb692ff0d92213c12f1" Feb 03 10:00:09 crc kubenswrapper[4756]: I0203 10:00:09.624671 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" path="/var/lib/kubelet/pods/7c3429c5-a971-48ef-a420-94f6f1717db8/volumes" Feb 03 10:00:13 crc kubenswrapper[4756]: I0203 10:00:13.614106 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:00:13 crc kubenswrapper[4756]: E0203 10:00:13.615010 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:00:24 crc kubenswrapper[4756]: I0203 10:00:24.615047 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:00:24 crc kubenswrapper[4756]: E0203 10:00:24.618038 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:00:31 crc kubenswrapper[4756]: I0203 10:00:31.767908 4756 scope.go:117] "RemoveContainer" containerID="bbdc33cc9fa87a5c29ad593bdcd5b4c97b1b35a614bfab26b39ccdb4ca22ab15" Feb 03 10:00:38 crc kubenswrapper[4756]: I0203 10:00:38.613994 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:00:38 crc kubenswrapper[4756]: E0203 10:00:38.614852 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:00:49 crc kubenswrapper[4756]: I0203 10:00:49.623107 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:00:49 crc kubenswrapper[4756]: E0203 10:00:49.624052 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.150866 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29501881-dszb5"] Feb 03 10:01:00 crc kubenswrapper[4756]: E0203 10:01:00.151919 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81294c99-0854-4af1-b552-41f9695aa7bc" containerName="collect-profiles" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.151937 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="81294c99-0854-4af1-b552-41f9695aa7bc" containerName="collect-profiles" Feb 03 10:01:00 crc kubenswrapper[4756]: E0203 10:01:00.151962 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerName="extract-content" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.151970 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerName="extract-content" Feb 03 10:01:00 crc kubenswrapper[4756]: E0203 10:01:00.151986 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerName="extract-utilities" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.151995 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerName="extract-utilities" Feb 03 10:01:00 crc kubenswrapper[4756]: E0203 10:01:00.152026 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerName="registry-server" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.152033 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerName="registry-server" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.152256 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3429c5-a971-48ef-a420-94f6f1717db8" containerName="registry-server" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.152267 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="81294c99-0854-4af1-b552-41f9695aa7bc" containerName="collect-profiles" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.152977 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.165179 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29501881-dszb5"] Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.211735 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-config-data\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.211844 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vzks\" (UniqueName: \"kubernetes.io/projected/bccf9d07-df25-4256-8a89-f2a47ee20a73-kube-api-access-8vzks\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.211976 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-fernet-keys\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.212079 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-combined-ca-bundle\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.314372 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-combined-ca-bundle\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.314475 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-config-data\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.314529 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vzks\" (UniqueName: \"kubernetes.io/projected/bccf9d07-df25-4256-8a89-f2a47ee20a73-kube-api-access-8vzks\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.314567 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-fernet-keys\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.321256 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-combined-ca-bundle\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.326672 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-config-data\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.326748 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-fernet-keys\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.340073 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vzks\" (UniqueName: \"kubernetes.io/projected/bccf9d07-df25-4256-8a89-f2a47ee20a73-kube-api-access-8vzks\") pod \"keystone-cron-29501881-dszb5\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.484261 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.614893 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:01:00 crc kubenswrapper[4756]: E0203 10:01:00.615770 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:01:00 crc kubenswrapper[4756]: I0203 10:01:00.931553 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29501881-dszb5"] Feb 03 10:01:01 crc kubenswrapper[4756]: I0203 10:01:01.640113 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29501881-dszb5" event={"ID":"bccf9d07-df25-4256-8a89-f2a47ee20a73","Type":"ContainerStarted","Data":"9bd49823ef14f0392b365fc041696e36f0d2c4efe6eca56db78f1660d207c8b0"} Feb 03 10:01:01 crc kubenswrapper[4756]: I0203 10:01:01.640169 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29501881-dszb5" event={"ID":"bccf9d07-df25-4256-8a89-f2a47ee20a73","Type":"ContainerStarted","Data":"ad3593151f0d92655322818d2f0b4a07a0469dd7525ea2ff07686dbedfe07d79"} Feb 03 10:01:01 crc kubenswrapper[4756]: I0203 10:01:01.658688 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29501881-dszb5" podStartSLOduration=1.658669419 podStartE2EDuration="1.658669419s" podCreationTimestamp="2026-02-03 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 10:01:01.650980448 +0000 UTC m=+3052.801447823" watchObservedRunningTime="2026-02-03 10:01:01.658669419 +0000 UTC m=+3052.809136794" Feb 03 10:01:03 crc kubenswrapper[4756]: I0203 10:01:03.647861 4756 generic.go:334] "Generic (PLEG): container finished" podID="bccf9d07-df25-4256-8a89-f2a47ee20a73" containerID="9bd49823ef14f0392b365fc041696e36f0d2c4efe6eca56db78f1660d207c8b0" exitCode=0 Feb 03 10:01:03 crc kubenswrapper[4756]: I0203 10:01:03.647961 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29501881-dszb5" event={"ID":"bccf9d07-df25-4256-8a89-f2a47ee20a73","Type":"ContainerDied","Data":"9bd49823ef14f0392b365fc041696e36f0d2c4efe6eca56db78f1660d207c8b0"} Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.085024 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.211091 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-combined-ca-bundle\") pod \"bccf9d07-df25-4256-8a89-f2a47ee20a73\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.211267 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-fernet-keys\") pod \"bccf9d07-df25-4256-8a89-f2a47ee20a73\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.211340 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vzks\" (UniqueName: \"kubernetes.io/projected/bccf9d07-df25-4256-8a89-f2a47ee20a73-kube-api-access-8vzks\") pod \"bccf9d07-df25-4256-8a89-f2a47ee20a73\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.211425 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-config-data\") pod \"bccf9d07-df25-4256-8a89-f2a47ee20a73\" (UID: \"bccf9d07-df25-4256-8a89-f2a47ee20a73\") " Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.217360 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bccf9d07-df25-4256-8a89-f2a47ee20a73-kube-api-access-8vzks" (OuterVolumeSpecName: "kube-api-access-8vzks") pod "bccf9d07-df25-4256-8a89-f2a47ee20a73" (UID: "bccf9d07-df25-4256-8a89-f2a47ee20a73"). InnerVolumeSpecName "kube-api-access-8vzks". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.217851 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bccf9d07-df25-4256-8a89-f2a47ee20a73" (UID: "bccf9d07-df25-4256-8a89-f2a47ee20a73"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.238913 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bccf9d07-df25-4256-8a89-f2a47ee20a73" (UID: "bccf9d07-df25-4256-8a89-f2a47ee20a73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.261712 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-config-data" (OuterVolumeSpecName: "config-data") pod "bccf9d07-df25-4256-8a89-f2a47ee20a73" (UID: "bccf9d07-df25-4256-8a89-f2a47ee20a73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.314048 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vzks\" (UniqueName: \"kubernetes.io/projected/bccf9d07-df25-4256-8a89-f2a47ee20a73-kube-api-access-8vzks\") on node \"crc\" DevicePath \"\"" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.314082 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.314094 4756 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.314104 4756 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bccf9d07-df25-4256-8a89-f2a47ee20a73-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.666251 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29501881-dszb5" event={"ID":"bccf9d07-df25-4256-8a89-f2a47ee20a73","Type":"ContainerDied","Data":"ad3593151f0d92655322818d2f0b4a07a0469dd7525ea2ff07686dbedfe07d79"} Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.666293 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad3593151f0d92655322818d2f0b4a07a0469dd7525ea2ff07686dbedfe07d79" Feb 03 10:01:05 crc kubenswrapper[4756]: I0203 10:01:05.666352 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29501881-dszb5" Feb 03 10:01:15 crc kubenswrapper[4756]: I0203 10:01:15.614027 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:01:15 crc kubenswrapper[4756]: E0203 10:01:15.614939 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:01:28 crc kubenswrapper[4756]: I0203 10:01:28.614565 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:01:28 crc kubenswrapper[4756]: E0203 10:01:28.615498 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.016040 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-98w8v"] Feb 03 10:01:32 crc kubenswrapper[4756]: E0203 10:01:32.017078 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bccf9d07-df25-4256-8a89-f2a47ee20a73" containerName="keystone-cron" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.017096 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="bccf9d07-df25-4256-8a89-f2a47ee20a73" containerName="keystone-cron" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.017538 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="bccf9d07-df25-4256-8a89-f2a47ee20a73" containerName="keystone-cron" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.020667 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.028882 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98w8v"] Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.126003 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-utilities\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.126217 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-catalog-content\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.126419 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mthx6\" (UniqueName: \"kubernetes.io/projected/1f796136-07a6-475b-a4e0-dc6b84d4b402-kube-api-access-mthx6\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.228819 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mthx6\" (UniqueName: \"kubernetes.io/projected/1f796136-07a6-475b-a4e0-dc6b84d4b402-kube-api-access-mthx6\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.228946 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-utilities\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.229020 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-catalog-content\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.229615 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-utilities\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.229643 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-catalog-content\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.260642 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mthx6\" (UniqueName: \"kubernetes.io/projected/1f796136-07a6-475b-a4e0-dc6b84d4b402-kube-api-access-mthx6\") pod \"certified-operators-98w8v\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.346255 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.896025 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-98w8v"] Feb 03 10:01:32 crc kubenswrapper[4756]: I0203 10:01:32.924232 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98w8v" event={"ID":"1f796136-07a6-475b-a4e0-dc6b84d4b402","Type":"ContainerStarted","Data":"05638b62c3af715042087768aeb56bab087dbbee88848ac12c4b953545282916"} Feb 03 10:01:33 crc kubenswrapper[4756]: I0203 10:01:33.933977 4756 generic.go:334] "Generic (PLEG): container finished" podID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerID="997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6" exitCode=0 Feb 03 10:01:33 crc kubenswrapper[4756]: I0203 10:01:33.934091 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98w8v" event={"ID":"1f796136-07a6-475b-a4e0-dc6b84d4b402","Type":"ContainerDied","Data":"997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6"} Feb 03 10:01:35 crc kubenswrapper[4756]: I0203 10:01:35.952717 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98w8v" event={"ID":"1f796136-07a6-475b-a4e0-dc6b84d4b402","Type":"ContainerStarted","Data":"7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e"} Feb 03 10:01:36 crc kubenswrapper[4756]: I0203 10:01:36.970299 4756 generic.go:334] "Generic (PLEG): container finished" podID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerID="7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e" exitCode=0 Feb 03 10:01:36 crc kubenswrapper[4756]: I0203 10:01:36.970403 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98w8v" event={"ID":"1f796136-07a6-475b-a4e0-dc6b84d4b402","Type":"ContainerDied","Data":"7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e"} Feb 03 10:01:37 crc kubenswrapper[4756]: I0203 10:01:37.980813 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98w8v" event={"ID":"1f796136-07a6-475b-a4e0-dc6b84d4b402","Type":"ContainerStarted","Data":"3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94"} Feb 03 10:01:38 crc kubenswrapper[4756]: I0203 10:01:38.003494 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-98w8v" podStartSLOduration=3.48708513 podStartE2EDuration="7.003475284s" podCreationTimestamp="2026-02-03 10:01:31 +0000 UTC" firstStartedPulling="2026-02-03 10:01:33.935927284 +0000 UTC m=+3085.086394659" lastFinishedPulling="2026-02-03 10:01:37.452317438 +0000 UTC m=+3088.602784813" observedRunningTime="2026-02-03 10:01:37.999650875 +0000 UTC m=+3089.150118260" watchObservedRunningTime="2026-02-03 10:01:38.003475284 +0000 UTC m=+3089.153942659" Feb 03 10:01:42 crc kubenswrapper[4756]: I0203 10:01:42.347019 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:42 crc kubenswrapper[4756]: I0203 10:01:42.347423 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:42 crc kubenswrapper[4756]: I0203 10:01:42.393827 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:42 crc kubenswrapper[4756]: I0203 10:01:42.613876 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:01:42 crc kubenswrapper[4756]: E0203 10:01:42.614373 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:01:43 crc kubenswrapper[4756]: I0203 10:01:43.104858 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:44 crc kubenswrapper[4756]: I0203 10:01:44.404115 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-98w8v"] Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.050944 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-98w8v" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerName="registry-server" containerID="cri-o://3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94" gracePeriod=2 Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.650924 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.827111 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mthx6\" (UniqueName: \"kubernetes.io/projected/1f796136-07a6-475b-a4e0-dc6b84d4b402-kube-api-access-mthx6\") pod \"1f796136-07a6-475b-a4e0-dc6b84d4b402\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.827217 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-utilities\") pod \"1f796136-07a6-475b-a4e0-dc6b84d4b402\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.827281 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-catalog-content\") pod \"1f796136-07a6-475b-a4e0-dc6b84d4b402\" (UID: \"1f796136-07a6-475b-a4e0-dc6b84d4b402\") " Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.828187 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-utilities" (OuterVolumeSpecName: "utilities") pod "1f796136-07a6-475b-a4e0-dc6b84d4b402" (UID: "1f796136-07a6-475b-a4e0-dc6b84d4b402"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.834726 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f796136-07a6-475b-a4e0-dc6b84d4b402-kube-api-access-mthx6" (OuterVolumeSpecName: "kube-api-access-mthx6") pod "1f796136-07a6-475b-a4e0-dc6b84d4b402" (UID: "1f796136-07a6-475b-a4e0-dc6b84d4b402"). InnerVolumeSpecName "kube-api-access-mthx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.877858 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f796136-07a6-475b-a4e0-dc6b84d4b402" (UID: "1f796136-07a6-475b-a4e0-dc6b84d4b402"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.929647 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mthx6\" (UniqueName: \"kubernetes.io/projected/1f796136-07a6-475b-a4e0-dc6b84d4b402-kube-api-access-mthx6\") on node \"crc\" DevicePath \"\"" Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.929703 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 10:01:45 crc kubenswrapper[4756]: I0203 10:01:45.929718 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f796136-07a6-475b-a4e0-dc6b84d4b402-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.063553 4756 generic.go:334] "Generic (PLEG): container finished" podID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerID="3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94" exitCode=0 Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.063620 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-98w8v" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.063625 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98w8v" event={"ID":"1f796136-07a6-475b-a4e0-dc6b84d4b402","Type":"ContainerDied","Data":"3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94"} Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.063937 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-98w8v" event={"ID":"1f796136-07a6-475b-a4e0-dc6b84d4b402","Type":"ContainerDied","Data":"05638b62c3af715042087768aeb56bab087dbbee88848ac12c4b953545282916"} Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.063982 4756 scope.go:117] "RemoveContainer" containerID="3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.085268 4756 scope.go:117] "RemoveContainer" containerID="7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.095863 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-98w8v"] Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.105506 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-98w8v"] Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.125559 4756 scope.go:117] "RemoveContainer" containerID="997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.162465 4756 scope.go:117] "RemoveContainer" containerID="3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94" Feb 03 10:01:46 crc kubenswrapper[4756]: E0203 10:01:46.163009 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94\": container with ID starting with 3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94 not found: ID does not exist" containerID="3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.163069 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94"} err="failed to get container status \"3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94\": rpc error: code = NotFound desc = could not find container \"3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94\": container with ID starting with 3b6e7ffbe8183af23ecec5ca8d2534d0c40bb106bdfb6f6a782aa6f3b34a9c94 not found: ID does not exist" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.163105 4756 scope.go:117] "RemoveContainer" containerID="7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e" Feb 03 10:01:46 crc kubenswrapper[4756]: E0203 10:01:46.163589 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e\": container with ID starting with 7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e not found: ID does not exist" containerID="7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.163634 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e"} err="failed to get container status \"7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e\": rpc error: code = NotFound desc = could not find container \"7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e\": container with ID starting with 7b207f08685918dd30d39e8d06db61cfebdc4d17e33524fc4bedb7b25c2e321e not found: ID does not exist" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.163660 4756 scope.go:117] "RemoveContainer" containerID="997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6" Feb 03 10:01:46 crc kubenswrapper[4756]: E0203 10:01:46.164688 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6\": container with ID starting with 997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6 not found: ID does not exist" containerID="997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6" Feb 03 10:01:46 crc kubenswrapper[4756]: I0203 10:01:46.164728 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6"} err="failed to get container status \"997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6\": rpc error: code = NotFound desc = could not find container \"997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6\": container with ID starting with 997d3706683220ebc4c86f12f6c118f694ff94c7e12368d635f78a65a85b0fa6 not found: ID does not exist" Feb 03 10:01:47 crc kubenswrapper[4756]: I0203 10:01:47.623077 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" path="/var/lib/kubelet/pods/1f796136-07a6-475b-a4e0-dc6b84d4b402/volumes" Feb 03 10:01:57 crc kubenswrapper[4756]: I0203 10:01:57.614137 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:01:57 crc kubenswrapper[4756]: E0203 10:01:57.615103 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:02:10 crc kubenswrapper[4756]: I0203 10:02:10.614028 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:02:10 crc kubenswrapper[4756]: E0203 10:02:10.614791 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:02:21 crc kubenswrapper[4756]: I0203 10:02:21.616102 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:02:21 crc kubenswrapper[4756]: E0203 10:02:21.617036 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:02:32 crc kubenswrapper[4756]: I0203 10:02:32.614219 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:02:32 crc kubenswrapper[4756]: E0203 10:02:32.615115 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:02:45 crc kubenswrapper[4756]: I0203 10:02:45.614110 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:02:45 crc kubenswrapper[4756]: E0203 10:02:45.631836 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:02:57 crc kubenswrapper[4756]: I0203 10:02:57.616373 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:02:57 crc kubenswrapper[4756]: E0203 10:02:57.617321 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:03:10 crc kubenswrapper[4756]: I0203 10:03:10.614133 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:03:10 crc kubenswrapper[4756]: E0203 10:03:10.615141 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:03:25 crc kubenswrapper[4756]: I0203 10:03:25.614396 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:03:25 crc kubenswrapper[4756]: I0203 10:03:25.949727 4756 generic.go:334] "Generic (PLEG): container finished" podID="c529c965-e4c3-4e32-a5b2-445aebef0f08" containerID="9141af1444872117c0828db828b928b1f2dab3fc997730d8a50dff377eaecefe" exitCode=0 Feb 03 10:03:25 crc kubenswrapper[4756]: I0203 10:03:25.949987 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c529c965-e4c3-4e32-a5b2-445aebef0f08","Type":"ContainerDied","Data":"9141af1444872117c0828db828b928b1f2dab3fc997730d8a50dff377eaecefe"} Feb 03 10:03:25 crc kubenswrapper[4756]: I0203 10:03:25.953658 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"e93ac21e9eba4fd7bad1c059d84ee34641b9f4a108f333d3d5f53ecefa6ab116"} Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.346228 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460274 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnhdq\" (UniqueName: \"kubernetes.io/projected/c529c965-e4c3-4e32-a5b2-445aebef0f08-kube-api-access-mnhdq\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460350 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-temporary\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460439 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-config-data\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460512 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ssh-key\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460540 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460574 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ca-certs\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460664 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config-secret\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460772 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.460819 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-workdir\") pod \"c529c965-e4c3-4e32-a5b2-445aebef0f08\" (UID: \"c529c965-e4c3-4e32-a5b2-445aebef0f08\") " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.461168 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.461327 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-config-data" (OuterVolumeSpecName: "config-data") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.466936 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c529c965-e4c3-4e32-a5b2-445aebef0f08-kube-api-access-mnhdq" (OuterVolumeSpecName: "kube-api-access-mnhdq") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "kube-api-access-mnhdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.467212 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.469065 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.487786 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.490601 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.493094 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.507753 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c529c965-e4c3-4e32-a5b2-445aebef0f08" (UID: "c529c965-e4c3-4e32-a5b2-445aebef0f08"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563535 4756 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563873 4756 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563894 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnhdq\" (UniqueName: \"kubernetes.io/projected/c529c965-e4c3-4e32-a5b2-445aebef0f08-kube-api-access-mnhdq\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563905 4756 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/c529c965-e4c3-4e32-a5b2-445aebef0f08-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563915 4756 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-config-data\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563923 4756 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563931 4756 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563939 4756 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.563947 4756 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c529c965-e4c3-4e32-a5b2-445aebef0f08-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.596664 4756 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.665162 4756 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.969679 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"c529c965-e4c3-4e32-a5b2-445aebef0f08","Type":"ContainerDied","Data":"2050a6e09bb8f32cfd7701c5bb271b4a0547f632780128ab2dc39b6786d070ca"} Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.969718 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2050a6e09bb8f32cfd7701c5bb271b4a0547f632780128ab2dc39b6786d070ca" Feb 03 10:03:27 crc kubenswrapper[4756]: I0203 10:03:27.969735 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.809723 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 03 10:03:35 crc kubenswrapper[4756]: E0203 10:03:35.810843 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerName="registry-server" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.810861 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerName="registry-server" Feb 03 10:03:35 crc kubenswrapper[4756]: E0203 10:03:35.810873 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c529c965-e4c3-4e32-a5b2-445aebef0f08" containerName="tempest-tests-tempest-tests-runner" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.810881 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c529c965-e4c3-4e32-a5b2-445aebef0f08" containerName="tempest-tests-tempest-tests-runner" Feb 03 10:03:35 crc kubenswrapper[4756]: E0203 10:03:35.810919 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerName="extract-content" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.810928 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerName="extract-content" Feb 03 10:03:35 crc kubenswrapper[4756]: E0203 10:03:35.810944 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerName="extract-utilities" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.810951 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerName="extract-utilities" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.811162 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f796136-07a6-475b-a4e0-dc6b84d4b402" containerName="registry-server" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.811180 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="c529c965-e4c3-4e32-a5b2-445aebef0f08" containerName="tempest-tests-tempest-tests-runner" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.811946 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.817066 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-m8rc4" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.836202 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.928485 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e5cc0402-4e5a-49ab-bd96-654e19fe4048\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:35 crc kubenswrapper[4756]: I0203 10:03:35.929755 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrhtg\" (UniqueName: \"kubernetes.io/projected/e5cc0402-4e5a-49ab-bd96-654e19fe4048-kube-api-access-rrhtg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e5cc0402-4e5a-49ab-bd96-654e19fe4048\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:36 crc kubenswrapper[4756]: I0203 10:03:36.034855 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e5cc0402-4e5a-49ab-bd96-654e19fe4048\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:36 crc kubenswrapper[4756]: I0203 10:03:36.035322 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrhtg\" (UniqueName: \"kubernetes.io/projected/e5cc0402-4e5a-49ab-bd96-654e19fe4048-kube-api-access-rrhtg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e5cc0402-4e5a-49ab-bd96-654e19fe4048\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:36 crc kubenswrapper[4756]: I0203 10:03:36.035370 4756 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e5cc0402-4e5a-49ab-bd96-654e19fe4048\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:36 crc kubenswrapper[4756]: I0203 10:03:36.066319 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrhtg\" (UniqueName: \"kubernetes.io/projected/e5cc0402-4e5a-49ab-bd96-654e19fe4048-kube-api-access-rrhtg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e5cc0402-4e5a-49ab-bd96-654e19fe4048\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:36 crc kubenswrapper[4756]: I0203 10:03:36.072166 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e5cc0402-4e5a-49ab-bd96-654e19fe4048\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:36 crc kubenswrapper[4756]: I0203 10:03:36.141634 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 03 10:03:36 crc kubenswrapper[4756]: I0203 10:03:36.609931 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 10:03:36 crc kubenswrapper[4756]: I0203 10:03:36.614550 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 03 10:03:37 crc kubenswrapper[4756]: I0203 10:03:37.041101 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e5cc0402-4e5a-49ab-bd96-654e19fe4048","Type":"ContainerStarted","Data":"c38dfef22a934042663855cc28a076598aa7d8dfb8db0e0a2062851437401293"} Feb 03 10:03:40 crc kubenswrapper[4756]: I0203 10:03:40.067698 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e5cc0402-4e5a-49ab-bd96-654e19fe4048","Type":"ContainerStarted","Data":"6299e1a39c07185dd93753af2465b4dd74acaca73d13782109ad133dbc717f1a"} Feb 03 10:03:40 crc kubenswrapper[4756]: I0203 10:03:40.090672 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.490743511 podStartE2EDuration="5.09065103s" podCreationTimestamp="2026-02-03 10:03:35 +0000 UTC" firstStartedPulling="2026-02-03 10:03:36.609266881 +0000 UTC m=+3207.759734266" lastFinishedPulling="2026-02-03 10:03:39.2091744 +0000 UTC m=+3210.359641785" observedRunningTime="2026-02-03 10:03:40.089854535 +0000 UTC m=+3211.240321910" watchObservedRunningTime="2026-02-03 10:03:40.09065103 +0000 UTC m=+3211.241118415" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.184869 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h5jjl/must-gather-ssccp"] Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.187091 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.205364 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h5jjl"/"kube-root-ca.crt" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.207857 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h5jjl"/"openshift-service-ca.crt" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.221012 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h5jjl/must-gather-ssccp"] Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.248527 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5a2ae189-29a7-41db-9916-910149f8b124-must-gather-output\") pod \"must-gather-ssccp\" (UID: \"5a2ae189-29a7-41db-9916-910149f8b124\") " pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.248618 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnrb4\" (UniqueName: \"kubernetes.io/projected/5a2ae189-29a7-41db-9916-910149f8b124-kube-api-access-hnrb4\") pod \"must-gather-ssccp\" (UID: \"5a2ae189-29a7-41db-9916-910149f8b124\") " pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.351834 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnrb4\" (UniqueName: \"kubernetes.io/projected/5a2ae189-29a7-41db-9916-910149f8b124-kube-api-access-hnrb4\") pod \"must-gather-ssccp\" (UID: \"5a2ae189-29a7-41db-9916-910149f8b124\") " pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.352154 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5a2ae189-29a7-41db-9916-910149f8b124-must-gather-output\") pod \"must-gather-ssccp\" (UID: \"5a2ae189-29a7-41db-9916-910149f8b124\") " pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.352615 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5a2ae189-29a7-41db-9916-910149f8b124-must-gather-output\") pod \"must-gather-ssccp\" (UID: \"5a2ae189-29a7-41db-9916-910149f8b124\") " pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.370672 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnrb4\" (UniqueName: \"kubernetes.io/projected/5a2ae189-29a7-41db-9916-910149f8b124-kube-api-access-hnrb4\") pod \"must-gather-ssccp\" (UID: \"5a2ae189-29a7-41db-9916-910149f8b124\") " pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:04:01 crc kubenswrapper[4756]: I0203 10:04:01.526681 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:04:02 crc kubenswrapper[4756]: I0203 10:04:02.003830 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h5jjl/must-gather-ssccp"] Feb 03 10:04:02 crc kubenswrapper[4756]: I0203 10:04:02.256125 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/must-gather-ssccp" event={"ID":"5a2ae189-29a7-41db-9916-910149f8b124","Type":"ContainerStarted","Data":"8afb457caedada59f5d409cc3e6f27b05516933573943380620b269fb6ae28a1"} Feb 03 10:04:07 crc kubenswrapper[4756]: I0203 10:04:07.762261 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/must-gather-ssccp" event={"ID":"5a2ae189-29a7-41db-9916-910149f8b124","Type":"ContainerStarted","Data":"322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8"} Feb 03 10:04:07 crc kubenswrapper[4756]: I0203 10:04:07.762871 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/must-gather-ssccp" event={"ID":"5a2ae189-29a7-41db-9916-910149f8b124","Type":"ContainerStarted","Data":"6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d"} Feb 03 10:04:07 crc kubenswrapper[4756]: I0203 10:04:07.780522 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h5jjl/must-gather-ssccp" podStartSLOduration=2.819252321 podStartE2EDuration="6.780506839s" podCreationTimestamp="2026-02-03 10:04:01 +0000 UTC" firstStartedPulling="2026-02-03 10:04:02.005723349 +0000 UTC m=+3233.156190724" lastFinishedPulling="2026-02-03 10:04:05.966977867 +0000 UTC m=+3237.117445242" observedRunningTime="2026-02-03 10:04:07.774263304 +0000 UTC m=+3238.924730689" watchObservedRunningTime="2026-02-03 10:04:07.780506839 +0000 UTC m=+3238.930974214" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.568435 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-r49w6"] Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.570329 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.572731 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-h5jjl"/"default-dockercfg-jnst4" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.672315 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-host\") pod \"crc-debug-r49w6\" (UID: \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\") " pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.672521 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlmb8\" (UniqueName: \"kubernetes.io/projected/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-kube-api-access-zlmb8\") pod \"crc-debug-r49w6\" (UID: \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\") " pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.774886 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-host\") pod \"crc-debug-r49w6\" (UID: \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\") " pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.775012 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlmb8\" (UniqueName: \"kubernetes.io/projected/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-kube-api-access-zlmb8\") pod \"crc-debug-r49w6\" (UID: \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\") " pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.775058 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-host\") pod \"crc-debug-r49w6\" (UID: \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\") " pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.801173 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlmb8\" (UniqueName: \"kubernetes.io/projected/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-kube-api-access-zlmb8\") pod \"crc-debug-r49w6\" (UID: \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\") " pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:04:12 crc kubenswrapper[4756]: I0203 10:04:12.889140 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:04:13 crc kubenswrapper[4756]: I0203 10:04:13.825795 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/crc-debug-r49w6" event={"ID":"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26","Type":"ContainerStarted","Data":"a206986ab5ad95f9c9d0b65553c34679c5c644a22c9def4febc48a601e12c460"} Feb 03 10:04:24 crc kubenswrapper[4756]: I0203 10:04:24.961164 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/crc-debug-r49w6" event={"ID":"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26","Type":"ContainerStarted","Data":"b64101f92ad09a0506e4e3e26cbfac81e9be9e0973570d654d1e4abffdfae6b9"} Feb 03 10:05:09 crc kubenswrapper[4756]: I0203 10:05:09.327102 4756 generic.go:334] "Generic (PLEG): container finished" podID="fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26" containerID="b64101f92ad09a0506e4e3e26cbfac81e9be9e0973570d654d1e4abffdfae6b9" exitCode=0 Feb 03 10:05:09 crc kubenswrapper[4756]: I0203 10:05:09.327193 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/crc-debug-r49w6" event={"ID":"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26","Type":"ContainerDied","Data":"b64101f92ad09a0506e4e3e26cbfac81e9be9e0973570d654d1e4abffdfae6b9"} Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.483792 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.525339 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-r49w6"] Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.534784 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-r49w6"] Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.559599 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlmb8\" (UniqueName: \"kubernetes.io/projected/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-kube-api-access-zlmb8\") pod \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\" (UID: \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\") " Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.559814 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-host\") pod \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\" (UID: \"fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26\") " Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.560578 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-host" (OuterVolumeSpecName: "host") pod "fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26" (UID: "fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.566186 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-kube-api-access-zlmb8" (OuterVolumeSpecName: "kube-api-access-zlmb8") pod "fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26" (UID: "fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26"). InnerVolumeSpecName "kube-api-access-zlmb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.662325 4756 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-host\") on node \"crc\" DevicePath \"\"" Feb 03 10:05:10 crc kubenswrapper[4756]: I0203 10:05:10.662371 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlmb8\" (UniqueName: \"kubernetes.io/projected/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26-kube-api-access-zlmb8\") on node \"crc\" DevicePath \"\"" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.345930 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a206986ab5ad95f9c9d0b65553c34679c5c644a22c9def4febc48a601e12c460" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.346036 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-r49w6" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.635191 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26" path="/var/lib/kubelet/pods/fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26/volumes" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.696711 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-x94p5"] Feb 03 10:05:11 crc kubenswrapper[4756]: E0203 10:05:11.697171 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26" containerName="container-00" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.697192 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26" containerName="container-00" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.697432 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf17dc4-ec7b-412c-93c9-5fa8b4f34c26" containerName="container-00" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.698126 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.700552 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-h5jjl"/"default-dockercfg-jnst4" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.783432 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-host\") pod \"crc-debug-x94p5\" (UID: \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\") " pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.783530 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfnhr\" (UniqueName: \"kubernetes.io/projected/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-kube-api-access-nfnhr\") pod \"crc-debug-x94p5\" (UID: \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\") " pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.884616 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfnhr\" (UniqueName: \"kubernetes.io/projected/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-kube-api-access-nfnhr\") pod \"crc-debug-x94p5\" (UID: \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\") " pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.884823 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-host\") pod \"crc-debug-x94p5\" (UID: \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\") " pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.884950 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-host\") pod \"crc-debug-x94p5\" (UID: \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\") " pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:11 crc kubenswrapper[4756]: I0203 10:05:11.901049 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfnhr\" (UniqueName: \"kubernetes.io/projected/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-kube-api-access-nfnhr\") pod \"crc-debug-x94p5\" (UID: \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\") " pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:12 crc kubenswrapper[4756]: I0203 10:05:12.022680 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:12 crc kubenswrapper[4756]: I0203 10:05:12.357370 4756 generic.go:334] "Generic (PLEG): container finished" podID="f6ab8670-0e68-4ce0-8077-d9cc9eacea82" containerID="03572fdc5d2a157546586a307d7d522bad8e167b2180d78807eb8948c4fd856d" exitCode=0 Feb 03 10:05:12 crc kubenswrapper[4756]: I0203 10:05:12.357654 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/crc-debug-x94p5" event={"ID":"f6ab8670-0e68-4ce0-8077-d9cc9eacea82","Type":"ContainerDied","Data":"03572fdc5d2a157546586a307d7d522bad8e167b2180d78807eb8948c4fd856d"} Feb 03 10:05:12 crc kubenswrapper[4756]: I0203 10:05:12.357686 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/crc-debug-x94p5" event={"ID":"f6ab8670-0e68-4ce0-8077-d9cc9eacea82","Type":"ContainerStarted","Data":"66cdb7c281b06c6feb85179fd881aca78c164f573f7300a270f522aa74c1b2ac"} Feb 03 10:05:12 crc kubenswrapper[4756]: I0203 10:05:12.811141 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-x94p5"] Feb 03 10:05:12 crc kubenswrapper[4756]: I0203 10:05:12.819877 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-x94p5"] Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.456083 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.516954 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-host\") pod \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\" (UID: \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\") " Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.517043 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfnhr\" (UniqueName: \"kubernetes.io/projected/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-kube-api-access-nfnhr\") pod \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\" (UID: \"f6ab8670-0e68-4ce0-8077-d9cc9eacea82\") " Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.517116 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-host" (OuterVolumeSpecName: "host") pod "f6ab8670-0e68-4ce0-8077-d9cc9eacea82" (UID: "f6ab8670-0e68-4ce0-8077-d9cc9eacea82"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.517920 4756 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-host\") on node \"crc\" DevicePath \"\"" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.524718 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-kube-api-access-nfnhr" (OuterVolumeSpecName: "kube-api-access-nfnhr") pod "f6ab8670-0e68-4ce0-8077-d9cc9eacea82" (UID: "f6ab8670-0e68-4ce0-8077-d9cc9eacea82"). InnerVolumeSpecName "kube-api-access-nfnhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.619326 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfnhr\" (UniqueName: \"kubernetes.io/projected/f6ab8670-0e68-4ce0-8077-d9cc9eacea82-kube-api-access-nfnhr\") on node \"crc\" DevicePath \"\"" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.626524 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6ab8670-0e68-4ce0-8077-d9cc9eacea82" path="/var/lib/kubelet/pods/f6ab8670-0e68-4ce0-8077-d9cc9eacea82/volumes" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.971180 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-rxmdj"] Feb 03 10:05:13 crc kubenswrapper[4756]: E0203 10:05:13.971632 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ab8670-0e68-4ce0-8077-d9cc9eacea82" containerName="container-00" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.971648 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ab8670-0e68-4ce0-8077-d9cc9eacea82" containerName="container-00" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.971880 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ab8670-0e68-4ce0-8077-d9cc9eacea82" containerName="container-00" Feb 03 10:05:13 crc kubenswrapper[4756]: I0203 10:05:13.972678 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.028881 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcjvv\" (UniqueName: \"kubernetes.io/projected/6873baad-280a-419c-bd2b-fcabe6e4a310-kube-api-access-rcjvv\") pod \"crc-debug-rxmdj\" (UID: \"6873baad-280a-419c-bd2b-fcabe6e4a310\") " pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.029206 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6873baad-280a-419c-bd2b-fcabe6e4a310-host\") pod \"crc-debug-rxmdj\" (UID: \"6873baad-280a-419c-bd2b-fcabe6e4a310\") " pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.131295 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6873baad-280a-419c-bd2b-fcabe6e4a310-host\") pod \"crc-debug-rxmdj\" (UID: \"6873baad-280a-419c-bd2b-fcabe6e4a310\") " pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.131377 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcjvv\" (UniqueName: \"kubernetes.io/projected/6873baad-280a-419c-bd2b-fcabe6e4a310-kube-api-access-rcjvv\") pod \"crc-debug-rxmdj\" (UID: \"6873baad-280a-419c-bd2b-fcabe6e4a310\") " pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.131416 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6873baad-280a-419c-bd2b-fcabe6e4a310-host\") pod \"crc-debug-rxmdj\" (UID: \"6873baad-280a-419c-bd2b-fcabe6e4a310\") " pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.152535 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcjvv\" (UniqueName: \"kubernetes.io/projected/6873baad-280a-419c-bd2b-fcabe6e4a310-kube-api-access-rcjvv\") pod \"crc-debug-rxmdj\" (UID: \"6873baad-280a-419c-bd2b-fcabe6e4a310\") " pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.295050 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:14 crc kubenswrapper[4756]: W0203 10:05:14.326290 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6873baad_280a_419c_bd2b_fcabe6e4a310.slice/crio-b92bfb422fdb93e20dd3090c0b39704ffb72f360d8824a8d7e194ee11db84dac WatchSource:0}: Error finding container b92bfb422fdb93e20dd3090c0b39704ffb72f360d8824a8d7e194ee11db84dac: Status 404 returned error can't find the container with id b92bfb422fdb93e20dd3090c0b39704ffb72f360d8824a8d7e194ee11db84dac Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.375629 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" event={"ID":"6873baad-280a-419c-bd2b-fcabe6e4a310","Type":"ContainerStarted","Data":"b92bfb422fdb93e20dd3090c0b39704ffb72f360d8824a8d7e194ee11db84dac"} Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.377556 4756 scope.go:117] "RemoveContainer" containerID="03572fdc5d2a157546586a307d7d522bad8e167b2180d78807eb8948c4fd856d" Feb 03 10:05:14 crc kubenswrapper[4756]: I0203 10:05:14.377598 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-x94p5" Feb 03 10:05:15 crc kubenswrapper[4756]: I0203 10:05:15.387719 4756 generic.go:334] "Generic (PLEG): container finished" podID="6873baad-280a-419c-bd2b-fcabe6e4a310" containerID="0765bdc46cd8670a8456bdd9d4ec1490e158e7b963f1e683c515751d4adc78c7" exitCode=0 Feb 03 10:05:15 crc kubenswrapper[4756]: I0203 10:05:15.387798 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" event={"ID":"6873baad-280a-419c-bd2b-fcabe6e4a310","Type":"ContainerDied","Data":"0765bdc46cd8670a8456bdd9d4ec1490e158e7b963f1e683c515751d4adc78c7"} Feb 03 10:05:15 crc kubenswrapper[4756]: I0203 10:05:15.447997 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-rxmdj"] Feb 03 10:05:15 crc kubenswrapper[4756]: I0203 10:05:15.458392 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h5jjl/crc-debug-rxmdj"] Feb 03 10:05:16 crc kubenswrapper[4756]: I0203 10:05:16.510300 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:16 crc kubenswrapper[4756]: I0203 10:05:16.577497 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6873baad-280a-419c-bd2b-fcabe6e4a310-host\") pod \"6873baad-280a-419c-bd2b-fcabe6e4a310\" (UID: \"6873baad-280a-419c-bd2b-fcabe6e4a310\") " Feb 03 10:05:16 crc kubenswrapper[4756]: I0203 10:05:16.577850 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcjvv\" (UniqueName: \"kubernetes.io/projected/6873baad-280a-419c-bd2b-fcabe6e4a310-kube-api-access-rcjvv\") pod \"6873baad-280a-419c-bd2b-fcabe6e4a310\" (UID: \"6873baad-280a-419c-bd2b-fcabe6e4a310\") " Feb 03 10:05:16 crc kubenswrapper[4756]: I0203 10:05:16.577612 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6873baad-280a-419c-bd2b-fcabe6e4a310-host" (OuterVolumeSpecName: "host") pod "6873baad-280a-419c-bd2b-fcabe6e4a310" (UID: "6873baad-280a-419c-bd2b-fcabe6e4a310"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 10:05:16 crc kubenswrapper[4756]: I0203 10:05:16.578424 4756 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6873baad-280a-419c-bd2b-fcabe6e4a310-host\") on node \"crc\" DevicePath \"\"" Feb 03 10:05:16 crc kubenswrapper[4756]: I0203 10:05:16.600226 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6873baad-280a-419c-bd2b-fcabe6e4a310-kube-api-access-rcjvv" (OuterVolumeSpecName: "kube-api-access-rcjvv") pod "6873baad-280a-419c-bd2b-fcabe6e4a310" (UID: "6873baad-280a-419c-bd2b-fcabe6e4a310"). InnerVolumeSpecName "kube-api-access-rcjvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:05:16 crc kubenswrapper[4756]: I0203 10:05:16.680556 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcjvv\" (UniqueName: \"kubernetes.io/projected/6873baad-280a-419c-bd2b-fcabe6e4a310-kube-api-access-rcjvv\") on node \"crc\" DevicePath \"\"" Feb 03 10:05:17 crc kubenswrapper[4756]: I0203 10:05:17.408398 4756 scope.go:117] "RemoveContainer" containerID="0765bdc46cd8670a8456bdd9d4ec1490e158e7b963f1e683c515751d4adc78c7" Feb 03 10:05:17 crc kubenswrapper[4756]: I0203 10:05:17.408604 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/crc-debug-rxmdj" Feb 03 10:05:17 crc kubenswrapper[4756]: I0203 10:05:17.627203 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6873baad-280a-419c-bd2b-fcabe6e4a310" path="/var/lib/kubelet/pods/6873baad-280a-419c-bd2b-fcabe6e4a310/volumes" Feb 03 10:05:31 crc kubenswrapper[4756]: I0203 10:05:31.435032 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f69df96f4-7db64_ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f/barbican-api/0.log" Feb 03 10:05:31 crc kubenswrapper[4756]: I0203 10:05:31.589754 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f69df96f4-7db64_ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f/barbican-api-log/0.log" Feb 03 10:05:31 crc kubenswrapper[4756]: I0203 10:05:31.628919 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5948f79bd-cl5cq_c34e66d1-0527-4f40-a676-1ac69c4d723f/barbican-keystone-listener/0.log" Feb 03 10:05:31 crc kubenswrapper[4756]: I0203 10:05:31.700593 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5948f79bd-cl5cq_c34e66d1-0527-4f40-a676-1ac69c4d723f/barbican-keystone-listener-log/0.log" Feb 03 10:05:31 crc kubenswrapper[4756]: I0203 10:05:31.820187 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68b68864ff-4znh2_8b049610-c8e6-40bd-98cb-138f42f1c35f/barbican-worker/0.log" Feb 03 10:05:31 crc kubenswrapper[4756]: I0203 10:05:31.889622 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68b68864ff-4znh2_8b049610-c8e6-40bd-98cb-138f42f1c35f/barbican-worker-log/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.032623 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks_5e81e9f8-2003-4c8b-8270-bf0214a074e7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.124972 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_198484fd-46b3-4299-8938-f49fdfdcacb0/ceilometer-central-agent/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.194101 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_198484fd-46b3-4299-8938-f49fdfdcacb0/ceilometer-notification-agent/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.264571 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_198484fd-46b3-4299-8938-f49fdfdcacb0/proxy-httpd/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.346797 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_198484fd-46b3-4299-8938-f49fdfdcacb0/sg-core/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.407893 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26/cinder-api/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.585585 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5abc83aa-9e60-4884-9633-d151104a151a/cinder-scheduler/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.602645 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26/cinder-api-log/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.644556 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5abc83aa-9e60-4884-9633-d151104a151a/probe/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.810204 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2_3c14e04f-9c29-45ae-84dc-bad543ff55d2/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:32 crc kubenswrapper[4756]: I0203 10:05:32.918517 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj_9851b3a8-4288-4c0b-9ae9-38a5c043682f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.069728 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-wtq86_2fa5e86f-f1ee-4e6e-9dba-26c7055a767e/init/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.217675 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-wtq86_2fa5e86f-f1ee-4e6e-9dba-26c7055a767e/init/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.287653 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-blt9q_2847ae70-333c-420d-b7a6-2c547477f35b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.319498 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-wtq86_2fa5e86f-f1ee-4e6e-9dba-26c7055a767e/dnsmasq-dns/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.462922 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2a921c4c-56ce-4689-b25d-0b4b69a01cdb/glance-httpd/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.484676 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2a921c4c-56ce-4689-b25d-0b4b69a01cdb/glance-log/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.677426 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bd8465a-e888-4b54-a570-76bcdc4736ac/glance-httpd/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.707360 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bd8465a-e888-4b54-a570-76bcdc4736ac/glance-log/0.log" Feb 03 10:05:33 crc kubenswrapper[4756]: I0203 10:05:33.934845 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d95f7b7bb-r6fd8_c34fbdee-8998-4e2d-bc98-c5c73ff35160/horizon/0.log" Feb 03 10:05:34 crc kubenswrapper[4756]: I0203 10:05:34.007725 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-9nclw_2396060c-bdc9-4f9c-8a42-c86984922418/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:34 crc kubenswrapper[4756]: I0203 10:05:34.159906 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d95f7b7bb-r6fd8_c34fbdee-8998-4e2d-bc98-c5c73ff35160/horizon-log/0.log" Feb 03 10:05:34 crc kubenswrapper[4756]: I0203 10:05:34.253395 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-q5fk9_bf98eb35-81c9-4cc5-9d5c-d81d97d716ad/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:34 crc kubenswrapper[4756]: I0203 10:05:34.461235 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29501881-dszb5_bccf9d07-df25-4256-8a89-f2a47ee20a73/keystone-cron/0.log" Feb 03 10:05:34 crc kubenswrapper[4756]: I0203 10:05:34.477937 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-765cf46959-4rcst_37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf/keystone-api/0.log" Feb 03 10:05:34 crc kubenswrapper[4756]: I0203 10:05:34.674339 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_dc0c1050-2088-4a2d-98bc-fb14078d6418/kube-state-metrics/0.log" Feb 03 10:05:34 crc kubenswrapper[4756]: I0203 10:05:34.761484 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m_9419b87e-a027-4476-a95b-e4bc884d90e7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:35 crc kubenswrapper[4756]: I0203 10:05:35.091617 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8547fc4b5c-w4sjb_3a5ca842-66f7-4ff0-b8f7-afe33276b34b/neutron-httpd/0.log" Feb 03 10:05:35 crc kubenswrapper[4756]: I0203 10:05:35.166222 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8547fc4b5c-w4sjb_3a5ca842-66f7-4ff0-b8f7-afe33276b34b/neutron-api/0.log" Feb 03 10:05:35 crc kubenswrapper[4756]: I0203 10:05:35.311950 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f_cebe42d5-5283-4c1e-a2f2-70db6d3e2257/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:35 crc kubenswrapper[4756]: I0203 10:05:35.854651 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ca27353e-8cea-4f67-a90b-9776e510e82e/nova-api-log/0.log" Feb 03 10:05:36 crc kubenswrapper[4756]: I0203 10:05:36.013409 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_84eaadd5-45a9-4055-a567-9e72fe3141b2/nova-cell0-conductor-conductor/0.log" Feb 03 10:05:36 crc kubenswrapper[4756]: I0203 10:05:36.059842 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ca27353e-8cea-4f67-a90b-9776e510e82e/nova-api-api/0.log" Feb 03 10:05:36 crc kubenswrapper[4756]: I0203 10:05:36.217867 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7978abb6-3f13-4b6e-8300-38a9d9cdd335/nova-cell1-conductor-conductor/0.log" Feb 03 10:05:36 crc kubenswrapper[4756]: I0203 10:05:36.323684 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_58183079-d9ac-4407-bec4-e93605aa884d/nova-cell1-novncproxy-novncproxy/0.log" Feb 03 10:05:36 crc kubenswrapper[4756]: I0203 10:05:36.540024 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-k5dsp_0573a518-65e1-4717-a24e-285e30cad247/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:36 crc kubenswrapper[4756]: I0203 10:05:36.692378 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_776f1390-6de5-4054-bff2-58621c198147/nova-metadata-log/0.log" Feb 03 10:05:36 crc kubenswrapper[4756]: I0203 10:05:36.937383 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_fead6bc0-9b80-4ab0-938c-19fad62f8178/nova-scheduler-scheduler/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.002104 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c/mysql-bootstrap/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.202219 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c/mysql-bootstrap/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.261711 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c/galera/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.435768 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_77baa89d-5260-4b1d-a633-a4f3e37ed6bb/mysql-bootstrap/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.691187 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_77baa89d-5260-4b1d-a633-a4f3e37ed6bb/galera/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.695352 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_77baa89d-5260-4b1d-a633-a4f3e37ed6bb/mysql-bootstrap/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.758492 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_776f1390-6de5-4054-bff2-58621c198147/nova-metadata-metadata/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.920690 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-d4749_f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a/openstack-network-exporter/0.log" Feb 03 10:05:37 crc kubenswrapper[4756]: I0203 10:05:37.992582 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3e4a2949-6bbf-4478-aee7-2cd98095c059/openstackclient/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.143483 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb566_4600dd3a-bdd0-44b1-8a6d-f34313d5df5d/ovsdb-server-init/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.371338 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb566_4600dd3a-bdd0-44b1-8a6d-f34313d5df5d/ovsdb-server/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.400392 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb566_4600dd3a-bdd0-44b1-8a6d-f34313d5df5d/ovs-vswitchd/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.409947 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb566_4600dd3a-bdd0-44b1-8a6d-f34313d5df5d/ovsdb-server-init/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.653203 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-wnf95_15695b6e-11fa-4913-b502-a78974ddf81c/ovn-controller/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.743537 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-jbd5l_3c6bade3-63d5-4342-8ded-feaf11b54263/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.874614 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_912cca3d-0179-4199-b236-a6ae37c9a80d/openstack-network-exporter/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.978671 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_912cca3d-0179-4199-b236-a6ae37c9a80d/ovn-northd/0.log" Feb 03 10:05:38 crc kubenswrapper[4756]: I0203 10:05:38.999101 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a2da6bae-fd2d-4f9c-b2aa-440d4264f17a/openstack-network-exporter/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.145767 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a2da6bae-fd2d-4f9c-b2aa-440d4264f17a/ovsdbserver-nb/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.319121 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bfba7eb1-da32-4bc6-8309-f076ad7b32a0/openstack-network-exporter/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.369617 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bfba7eb1-da32-4bc6-8309-f076ad7b32a0/ovsdbserver-sb/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.634009 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f5fdf688d-jsswq_5446426b-adaf-4af7-97f8-b8b2d798d364/placement-log/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.668809 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60b10ab9-d28a-4c83-9e67-02d1499685ce/setup-container/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.671675 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f5fdf688d-jsswq_5446426b-adaf-4af7-97f8-b8b2d798d364/placement-api/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.926591 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60b10ab9-d28a-4c83-9e67-02d1499685ce/setup-container/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.938313 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60b10ab9-d28a-4c83-9e67-02d1499685ce/rabbitmq/0.log" Feb 03 10:05:39 crc kubenswrapper[4756]: I0203 10:05:39.966142 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f76f412-ff7c-4857-9670-4721dc5c3080/setup-container/0.log" Feb 03 10:05:40 crc kubenswrapper[4756]: I0203 10:05:40.204286 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f76f412-ff7c-4857-9670-4721dc5c3080/rabbitmq/0.log" Feb 03 10:05:40 crc kubenswrapper[4756]: I0203 10:05:40.236360 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f76f412-ff7c-4857-9670-4721dc5c3080/setup-container/0.log" Feb 03 10:05:40 crc kubenswrapper[4756]: I0203 10:05:40.278091 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw_4ae5e9e1-d148-4487-862a-b2312222f2b0/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:40 crc kubenswrapper[4756]: I0203 10:05:40.467994 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-xvkxf_ce26979b-e0a6-41a5-b56d-e1800d5d6704/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:40 crc kubenswrapper[4756]: I0203 10:05:40.544716 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v_200f01ff-d33a-474b-944e-e6ed2486db68/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:40 crc kubenswrapper[4756]: I0203 10:05:40.687165 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9g7tk_5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:40 crc kubenswrapper[4756]: I0203 10:05:40.804568 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-7jsx4_848ce5a7-c89c-4cac-8275-263ca14eac78/ssh-known-hosts-edpm-deployment/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.004684 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85ff9cc7c5-ns7sz_74be27bd-98ba-4f2c-bd2e-ea5855812a55/proxy-server/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.081951 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85ff9cc7c5-ns7sz_74be27bd-98ba-4f2c-bd2e-ea5855812a55/proxy-httpd/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.316802 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-pqfmb_f356b5c0-fae0-4a67-9624-4fbe59bd2f9e/swift-ring-rebalance/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.452514 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/account-auditor/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.519201 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/account-reaper/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.542743 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/account-replicator/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.662417 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/container-auditor/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.680043 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/account-server/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.796173 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/container-replicator/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.809798 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/container-server/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.926050 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-auditor/0.log" Feb 03 10:05:41 crc kubenswrapper[4756]: I0203 10:05:41.956535 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/container-updater/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.051325 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-expirer/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.083609 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-replicator/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.155223 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-server/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.167427 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-updater/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.334561 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/swift-recon-cron/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.352253 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/rsync/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.532877 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd_17726fdd-5344-437e-8060-5e26bcfb28cd/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.690264 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c529c965-e4c3-4e32-a5b2-445aebef0f08/tempest-tests-tempest-tests-runner/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.750180 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_e5cc0402-4e5a-49ab-bd96-654e19fe4048/test-operator-logs-container/0.log" Feb 03 10:05:42 crc kubenswrapper[4756]: I0203 10:05:42.889107 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc_4a58ee0e-41e0-4040-99c1-3261c3416a29/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:05:43 crc kubenswrapper[4756]: I0203 10:05:43.565635 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:05:43 crc kubenswrapper[4756]: I0203 10:05:43.566052 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:05:51 crc kubenswrapper[4756]: I0203 10:05:51.316695 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ce92a835-f667-4f5b-b9e1-c8afcf476b9a/memcached/0.log" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.468836 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x9ts2"] Feb 03 10:05:58 crc kubenswrapper[4756]: E0203 10:05:58.471536 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6873baad-280a-419c-bd2b-fcabe6e4a310" containerName="container-00" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.471643 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6873baad-280a-419c-bd2b-fcabe6e4a310" containerName="container-00" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.471955 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="6873baad-280a-419c-bd2b-fcabe6e4a310" containerName="container-00" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.473923 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.503357 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9ts2"] Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.564724 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-catalog-content\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.564817 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-utilities\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.564843 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrhw8\" (UniqueName: \"kubernetes.io/projected/8e7584b9-1503-4fe2-8d52-de6d083285e4-kube-api-access-zrhw8\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.666963 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-catalog-content\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.667120 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-utilities\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.667153 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrhw8\" (UniqueName: \"kubernetes.io/projected/8e7584b9-1503-4fe2-8d52-de6d083285e4-kube-api-access-zrhw8\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.667408 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-catalog-content\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.667857 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-utilities\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.690671 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrhw8\" (UniqueName: \"kubernetes.io/projected/8e7584b9-1503-4fe2-8d52-de6d083285e4-kube-api-access-zrhw8\") pod \"redhat-operators-x9ts2\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:58 crc kubenswrapper[4756]: I0203 10:05:58.796955 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:05:59 crc kubenswrapper[4756]: I0203 10:05:59.446606 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9ts2"] Feb 03 10:05:59 crc kubenswrapper[4756]: I0203 10:05:59.805823 4756 generic.go:334] "Generic (PLEG): container finished" podID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerID="c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5" exitCode=0 Feb 03 10:05:59 crc kubenswrapper[4756]: I0203 10:05:59.805932 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9ts2" event={"ID":"8e7584b9-1503-4fe2-8d52-de6d083285e4","Type":"ContainerDied","Data":"c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5"} Feb 03 10:05:59 crc kubenswrapper[4756]: I0203 10:05:59.806152 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9ts2" event={"ID":"8e7584b9-1503-4fe2-8d52-de6d083285e4","Type":"ContainerStarted","Data":"a63a3f87713e340b5bb4e14e052d22bf8fb09002e1210c9679d9371eebee3db5"} Feb 03 10:06:00 crc kubenswrapper[4756]: I0203 10:06:00.816933 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9ts2" event={"ID":"8e7584b9-1503-4fe2-8d52-de6d083285e4","Type":"ContainerStarted","Data":"c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea"} Feb 03 10:06:04 crc kubenswrapper[4756]: I0203 10:06:04.850293 4756 generic.go:334] "Generic (PLEG): container finished" podID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerID="c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea" exitCode=0 Feb 03 10:06:04 crc kubenswrapper[4756]: I0203 10:06:04.850387 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9ts2" event={"ID":"8e7584b9-1503-4fe2-8d52-de6d083285e4","Type":"ContainerDied","Data":"c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea"} Feb 03 10:06:05 crc kubenswrapper[4756]: I0203 10:06:05.860851 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9ts2" event={"ID":"8e7584b9-1503-4fe2-8d52-de6d083285e4","Type":"ContainerStarted","Data":"b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9"} Feb 03 10:06:05 crc kubenswrapper[4756]: I0203 10:06:05.889140 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x9ts2" podStartSLOduration=2.382787954 podStartE2EDuration="7.889120862s" podCreationTimestamp="2026-02-03 10:05:58 +0000 UTC" firstStartedPulling="2026-02-03 10:05:59.807559426 +0000 UTC m=+3350.958026791" lastFinishedPulling="2026-02-03 10:06:05.313892324 +0000 UTC m=+3356.464359699" observedRunningTime="2026-02-03 10:06:05.882838366 +0000 UTC m=+3357.033305751" watchObservedRunningTime="2026-02-03 10:06:05.889120862 +0000 UTC m=+3357.039588247" Feb 03 10:06:08 crc kubenswrapper[4756]: I0203 10:06:08.797097 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:06:08 crc kubenswrapper[4756]: I0203 10:06:08.797665 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:06:09 crc kubenswrapper[4756]: I0203 10:06:09.862526 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x9ts2" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="registry-server" probeResult="failure" output=< Feb 03 10:06:09 crc kubenswrapper[4756]: timeout: failed to connect service ":50051" within 1s Feb 03 10:06:09 crc kubenswrapper[4756]: > Feb 03 10:06:11 crc kubenswrapper[4756]: I0203 10:06:11.862409 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/util/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.058363 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/pull/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.109704 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/util/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.129743 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/pull/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.306884 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/util/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.321153 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/pull/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.372523 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/extract/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.626887 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b6c4d8c5f-6pbkt_a6153ce6-006d-4351-8a9b-aa0abe1cbdad/manager/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.685764 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d874c8fc-dplk2_cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce/manager/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.793010 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d9697b7f4-qnklz_d74f0516-3682-4b80-beb4-f7237c4daff8/manager/0.log" Feb 03 10:06:12 crc kubenswrapper[4756]: I0203 10:06:12.956192 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8886f4c47-pgdzr_6dd1801b-ac57-4d28-9554-ef8ad7148c08/manager/0.log" Feb 03 10:06:13 crc kubenswrapper[4756]: I0203 10:06:13.247494 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69d6db494d-564mj_63c547ee-bb8c-4f1a-b29a-12a80c1749f4/manager/0.log" Feb 03 10:06:13 crc kubenswrapper[4756]: I0203 10:06:13.367693 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5fb775575f-hvnx5_d9f68e89-0c29-4d3d-9dae-7448ab35206f/manager/0.log" Feb 03 10:06:13 crc kubenswrapper[4756]: I0203 10:06:13.577854 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:06:13 crc kubenswrapper[4756]: I0203 10:06:13.577924 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:06:13 crc kubenswrapper[4756]: I0203 10:06:13.588993 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f4b8bd54d-zhlxl_ad043c5b-465c-43b2-b73c-45e605fc7b25/manager/0.log" Feb 03 10:06:13 crc kubenswrapper[4756]: I0203 10:06:13.764305 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79955696d6-vvkmq_aae19bbd-e8ec-4a6f-8e77-9dd46373e31b/manager/0.log" Feb 03 10:06:13 crc kubenswrapper[4756]: I0203 10:06:13.955524 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-84f48565d4-w7vdf_919fda51-9a13-4233-ba62-a85ac3559a02/manager/0.log" Feb 03 10:06:14 crc kubenswrapper[4756]: I0203 10:06:14.017419 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7dd968899f-jwlpw_d8ad4baf-2108-4286-a2f6-5a8ab537185c/manager/0.log" Feb 03 10:06:14 crc kubenswrapper[4756]: I0203 10:06:14.188493 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf948998-5q5s7_aacb5f03-cd60-4589-9758-9fb4e1b68030/manager/0.log" Feb 03 10:06:14 crc kubenswrapper[4756]: I0203 10:06:14.380638 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-585dbc889-gn2d2_4483922a-5bb6-40da-bb78-cdeca488ccd8/manager/0.log" Feb 03 10:06:14 crc kubenswrapper[4756]: I0203 10:06:14.539015 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-55bff696bd-qnwj2_999c5dd9-6dbb-44b9-bdd7-52c146d34e87/manager/0.log" Feb 03 10:06:14 crc kubenswrapper[4756]: I0203 10:06:14.603058 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6687f8d877-bqthj_932503a4-3932-4127-a2b2-a2e0291782b7/manager/0.log" Feb 03 10:06:14 crc kubenswrapper[4756]: I0203 10:06:14.797745 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx_a0ebd520-4c15-4c12-89fe-2eb8a407ca25/manager/0.log" Feb 03 10:06:14 crc kubenswrapper[4756]: I0203 10:06:14.999074 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-7f67889c79-dwpgg_2e4a4670-318d-4489-9a71-a084e2445361/operator/0.log" Feb 03 10:06:15 crc kubenswrapper[4756]: I0203 10:06:15.201478 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-r84l6_9dc52ada-64d7-45a9-b880-ded26c6ec2f6/registry-server/0.log" Feb 03 10:06:15 crc kubenswrapper[4756]: I0203 10:06:15.395344 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-788c46999f-wl9ft_bd1225c5-fbb5-40f5-8e71-6c1075feeb16/manager/0.log" Feb 03 10:06:15 crc kubenswrapper[4756]: I0203 10:06:15.649162 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b964cf4cd-7rrf6_58c01e66-fe0e-45f2-8caf-df41e1dad607/manager/0.log" Feb 03 10:06:15 crc kubenswrapper[4756]: I0203 10:06:15.693999 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-lqghx_9fe0931c-fae1-4dfd-968d-bccf61682fa5/operator/0.log" Feb 03 10:06:15 crc kubenswrapper[4756]: I0203 10:06:15.957042 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68fc8c869-w87dt_51a902ff-23e6-4578-809f-0de9a6c5cbae/manager/0.log" Feb 03 10:06:16 crc kubenswrapper[4756]: I0203 10:06:16.372263 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c6ffddb96-rvk7w_2183ff26-c37d-472e-bf53-e0beba79a31d/manager/0.log" Feb 03 10:06:16 crc kubenswrapper[4756]: I0203 10:06:16.376256 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64b5b76f97-2cw2v_d7f70a3a-d4db-4d12-8838-9d8923f06213/manager/0.log" Feb 03 10:06:16 crc kubenswrapper[4756]: I0203 10:06:16.378515 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56f8bfcd9f-l5wms_04edde9e-f443-4e9e-929b-65e2d9258ff2/manager/0.log" Feb 03 10:06:16 crc kubenswrapper[4756]: I0203 10:06:16.518267 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-9rdzw_ef6310fb-d44c-449f-b3d0-72c20e4b3d7b/manager/0.log" Feb 03 10:06:18 crc kubenswrapper[4756]: I0203 10:06:18.856050 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:06:18 crc kubenswrapper[4756]: I0203 10:06:18.907105 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:06:19 crc kubenswrapper[4756]: I0203 10:06:19.092563 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9ts2"] Feb 03 10:06:19 crc kubenswrapper[4756]: I0203 10:06:19.971555 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x9ts2" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="registry-server" containerID="cri-o://b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9" gracePeriod=2 Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.603888 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.764024 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-catalog-content\") pod \"8e7584b9-1503-4fe2-8d52-de6d083285e4\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.764213 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-utilities\") pod \"8e7584b9-1503-4fe2-8d52-de6d083285e4\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.764363 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrhw8\" (UniqueName: \"kubernetes.io/projected/8e7584b9-1503-4fe2-8d52-de6d083285e4-kube-api-access-zrhw8\") pod \"8e7584b9-1503-4fe2-8d52-de6d083285e4\" (UID: \"8e7584b9-1503-4fe2-8d52-de6d083285e4\") " Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.765465 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-utilities" (OuterVolumeSpecName: "utilities") pod "8e7584b9-1503-4fe2-8d52-de6d083285e4" (UID: "8e7584b9-1503-4fe2-8d52-de6d083285e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.777837 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e7584b9-1503-4fe2-8d52-de6d083285e4-kube-api-access-zrhw8" (OuterVolumeSpecName: "kube-api-access-zrhw8") pod "8e7584b9-1503-4fe2-8d52-de6d083285e4" (UID: "8e7584b9-1503-4fe2-8d52-de6d083285e4"). InnerVolumeSpecName "kube-api-access-zrhw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.868821 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.868868 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrhw8\" (UniqueName: \"kubernetes.io/projected/8e7584b9-1503-4fe2-8d52-de6d083285e4-kube-api-access-zrhw8\") on node \"crc\" DevicePath \"\"" Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.877303 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e7584b9-1503-4fe2-8d52-de6d083285e4" (UID: "8e7584b9-1503-4fe2-8d52-de6d083285e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.971416 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7584b9-1503-4fe2-8d52-de6d083285e4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.983026 4756 generic.go:334] "Generic (PLEG): container finished" podID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerID="b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9" exitCode=0 Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.983080 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9ts2" event={"ID":"8e7584b9-1503-4fe2-8d52-de6d083285e4","Type":"ContainerDied","Data":"b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9"} Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.983103 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9ts2" Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.983129 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9ts2" event={"ID":"8e7584b9-1503-4fe2-8d52-de6d083285e4","Type":"ContainerDied","Data":"a63a3f87713e340b5bb4e14e052d22bf8fb09002e1210c9679d9371eebee3db5"} Feb 03 10:06:20 crc kubenswrapper[4756]: I0203 10:06:20.983163 4756 scope.go:117] "RemoveContainer" containerID="b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.005786 4756 scope.go:117] "RemoveContainer" containerID="c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.035148 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9ts2"] Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.046765 4756 scope.go:117] "RemoveContainer" containerID="c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.053843 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x9ts2"] Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.080104 4756 scope.go:117] "RemoveContainer" containerID="b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9" Feb 03 10:06:21 crc kubenswrapper[4756]: E0203 10:06:21.080559 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9\": container with ID starting with b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9 not found: ID does not exist" containerID="b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.080608 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9"} err="failed to get container status \"b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9\": rpc error: code = NotFound desc = could not find container \"b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9\": container with ID starting with b45a69e8f65bed4f11456a28baf60833e2f1ea1b99a011a015ad8997675469b9 not found: ID does not exist" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.080638 4756 scope.go:117] "RemoveContainer" containerID="c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea" Feb 03 10:06:21 crc kubenswrapper[4756]: E0203 10:06:21.081240 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea\": container with ID starting with c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea not found: ID does not exist" containerID="c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.081291 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea"} err="failed to get container status \"c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea\": rpc error: code = NotFound desc = could not find container \"c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea\": container with ID starting with c4ea9cfb01c234298f4d7140fb05e184516d555e6ec0c573e006d3ded33ce8ea not found: ID does not exist" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.081319 4756 scope.go:117] "RemoveContainer" containerID="c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5" Feb 03 10:06:21 crc kubenswrapper[4756]: E0203 10:06:21.081633 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5\": container with ID starting with c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5 not found: ID does not exist" containerID="c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.081665 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5"} err="failed to get container status \"c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5\": rpc error: code = NotFound desc = could not find container \"c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5\": container with ID starting with c6498ad0875be33d4aa1e6df8b5e7c7fa7eaa11f422130652457cf1f51f764d5 not found: ID does not exist" Feb 03 10:06:21 crc kubenswrapper[4756]: I0203 10:06:21.625277 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" path="/var/lib/kubelet/pods/8e7584b9-1503-4fe2-8d52-de6d083285e4/volumes" Feb 03 10:06:36 crc kubenswrapper[4756]: I0203 10:06:36.976514 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qzff8_641583b7-728e-4e81-ad79-efcc524c5a10/control-plane-machine-set-operator/0.log" Feb 03 10:06:37 crc kubenswrapper[4756]: I0203 10:06:37.160390 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bk8nb_f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e/kube-rbac-proxy/0.log" Feb 03 10:06:37 crc kubenswrapper[4756]: I0203 10:06:37.172819 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bk8nb_f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e/machine-api-operator/0.log" Feb 03 10:06:43 crc kubenswrapper[4756]: I0203 10:06:43.566364 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:06:43 crc kubenswrapper[4756]: I0203 10:06:43.567367 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:06:43 crc kubenswrapper[4756]: I0203 10:06:43.567466 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 10:06:43 crc kubenswrapper[4756]: I0203 10:06:43.568683 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e93ac21e9eba4fd7bad1c059d84ee34641b9f4a108f333d3d5f53ecefa6ab116"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 10:06:43 crc kubenswrapper[4756]: I0203 10:06:43.568758 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://e93ac21e9eba4fd7bad1c059d84ee34641b9f4a108f333d3d5f53ecefa6ab116" gracePeriod=600 Feb 03 10:06:44 crc kubenswrapper[4756]: I0203 10:06:44.186492 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="e93ac21e9eba4fd7bad1c059d84ee34641b9f4a108f333d3d5f53ecefa6ab116" exitCode=0 Feb 03 10:06:44 crc kubenswrapper[4756]: I0203 10:06:44.186572 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"e93ac21e9eba4fd7bad1c059d84ee34641b9f4a108f333d3d5f53ecefa6ab116"} Feb 03 10:06:44 crc kubenswrapper[4756]: I0203 10:06:44.186828 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3"} Feb 03 10:06:44 crc kubenswrapper[4756]: I0203 10:06:44.186848 4756 scope.go:117] "RemoveContainer" containerID="252466476d36c96313f5029ae04e7f05126c6e7dc09e3bed17b634c4c4ea968b" Feb 03 10:06:50 crc kubenswrapper[4756]: I0203 10:06:50.581718 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-fm5vv_e8c43b40-0d26-4582-8ffd-b2387527edd6/cert-manager-controller/0.log" Feb 03 10:06:50 crc kubenswrapper[4756]: I0203 10:06:50.846961 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-gnvfj_2c537fcd-648d-4fea-ab9a-0e0edfa8b14b/cert-manager-cainjector/0.log" Feb 03 10:06:50 crc kubenswrapper[4756]: I0203 10:06:50.874317 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-gnr5k_a4b8e083-79fc-4bbe-b3c8-16e237b63458/cert-manager-webhook/0.log" Feb 03 10:07:04 crc kubenswrapper[4756]: I0203 10:07:04.672058 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-zscxn_ac470c2b-de36-4daa-bce0-8163e27e43e7/nmstate-console-plugin/0.log" Feb 03 10:07:04 crc kubenswrapper[4756]: I0203 10:07:04.945844 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-75x2d_63b70a58-8d39-418a-bedc-93f9a94540a4/kube-rbac-proxy/0.log" Feb 03 10:07:04 crc kubenswrapper[4756]: I0203 10:07:04.961349 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fv282_e827e9ff-5a84-42cc-9927-56904973c3ad/nmstate-handler/0.log" Feb 03 10:07:05 crc kubenswrapper[4756]: I0203 10:07:05.146486 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-75x2d_63b70a58-8d39-418a-bedc-93f9a94540a4/nmstate-metrics/0.log" Feb 03 10:07:05 crc kubenswrapper[4756]: I0203 10:07:05.227797 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-wvckj_6374dff8-08c2-49fe-b5f4-0eb5142f61bf/nmstate-operator/0.log" Feb 03 10:07:05 crc kubenswrapper[4756]: I0203 10:07:05.384665 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-cjp9b_b9dfc50f-d7ab-4806-aac1-70e36788c60e/nmstate-webhook/0.log" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.128436 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pn9kj"] Feb 03 10:07:29 crc kubenswrapper[4756]: E0203 10:07:29.129505 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="extract-content" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.129522 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="extract-content" Feb 03 10:07:29 crc kubenswrapper[4756]: E0203 10:07:29.129564 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="registry-server" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.129572 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="registry-server" Feb 03 10:07:29 crc kubenswrapper[4756]: E0203 10:07:29.129597 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="extract-utilities" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.129605 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="extract-utilities" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.129828 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e7584b9-1503-4fe2-8d52-de6d083285e4" containerName="registry-server" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.131631 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.148487 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pn9kj"] Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.223062 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-catalog-content\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.223115 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-utilities\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.223180 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m56ql\" (UniqueName: \"kubernetes.io/projected/06a32867-0d73-4505-a0df-cf49e30369fa-kube-api-access-m56ql\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.324855 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-catalog-content\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.324907 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-utilities\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.324983 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m56ql\" (UniqueName: \"kubernetes.io/projected/06a32867-0d73-4505-a0df-cf49e30369fa-kube-api-access-m56ql\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.325350 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-catalog-content\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.325393 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-utilities\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.346761 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m56ql\" (UniqueName: \"kubernetes.io/projected/06a32867-0d73-4505-a0df-cf49e30369fa-kube-api-access-m56ql\") pod \"community-operators-pn9kj\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:29 crc kubenswrapper[4756]: I0203 10:07:29.452624 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:30 crc kubenswrapper[4756]: I0203 10:07:30.055858 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pn9kj"] Feb 03 10:07:30 crc kubenswrapper[4756]: I0203 10:07:30.585383 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-dc6ld_16bed8ad-994f-4826-b43d-bc8ef79bce7e/kube-rbac-proxy/0.log" Feb 03 10:07:30 crc kubenswrapper[4756]: I0203 10:07:30.606905 4756 generic.go:334] "Generic (PLEG): container finished" podID="06a32867-0d73-4505-a0df-cf49e30369fa" containerID="69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a" exitCode=0 Feb 03 10:07:30 crc kubenswrapper[4756]: I0203 10:07:30.607184 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn9kj" event={"ID":"06a32867-0d73-4505-a0df-cf49e30369fa","Type":"ContainerDied","Data":"69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a"} Feb 03 10:07:30 crc kubenswrapper[4756]: I0203 10:07:30.607309 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn9kj" event={"ID":"06a32867-0d73-4505-a0df-cf49e30369fa","Type":"ContainerStarted","Data":"c4bb47f2999ed9e30474a5a03274b3c67f6f8710606be49f6a02090e61719d03"} Feb 03 10:07:30 crc kubenswrapper[4756]: I0203 10:07:30.698063 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-dc6ld_16bed8ad-994f-4826-b43d-bc8ef79bce7e/controller/0.log" Feb 03 10:07:30 crc kubenswrapper[4756]: I0203 10:07:30.862846 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-frr-files/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.029776 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-frr-files/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.061908 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-metrics/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.072156 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-reloader/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.080045 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-reloader/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.270398 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-frr-files/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.270538 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-metrics/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.317341 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-reloader/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.317664 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-metrics/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.521735 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-metrics/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.530476 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-reloader/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.569954 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-frr-files/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.590754 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/controller/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.645320 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn9kj" event={"ID":"06a32867-0d73-4505-a0df-cf49e30369fa","Type":"ContainerStarted","Data":"e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72"} Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.732276 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/frr-metrics/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.797062 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/kube-rbac-proxy/0.log" Feb 03 10:07:31 crc kubenswrapper[4756]: I0203 10:07:31.842074 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/kube-rbac-proxy-frr/0.log" Feb 03 10:07:32 crc kubenswrapper[4756]: I0203 10:07:32.059213 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/reloader/0.log" Feb 03 10:07:32 crc kubenswrapper[4756]: I0203 10:07:32.076582 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-bwmnx_bcaa1bc4-f6a5-4216-8581-13cbf07c2504/frr-k8s-webhook-server/0.log" Feb 03 10:07:32 crc kubenswrapper[4756]: I0203 10:07:32.319847 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-576684d48c-mkg4k_fc59b7e7-12f3-4501-94f9-07f3338d7835/manager/0.log" Feb 03 10:07:32 crc kubenswrapper[4756]: I0203 10:07:32.578527 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-87484dcc8-cbd6x_12aaddd9-3223-4da5-8040-4149116fede4/webhook-server/0.log" Feb 03 10:07:32 crc kubenswrapper[4756]: I0203 10:07:32.610667 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-44db6_98d5a2ec-16f8-4654-b5ba-0c32b08d0fda/kube-rbac-proxy/0.log" Feb 03 10:07:32 crc kubenswrapper[4756]: I0203 10:07:32.656183 4756 generic.go:334] "Generic (PLEG): container finished" podID="06a32867-0d73-4505-a0df-cf49e30369fa" containerID="e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72" exitCode=0 Feb 03 10:07:32 crc kubenswrapper[4756]: I0203 10:07:32.656237 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn9kj" event={"ID":"06a32867-0d73-4505-a0df-cf49e30369fa","Type":"ContainerDied","Data":"e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72"} Feb 03 10:07:33 crc kubenswrapper[4756]: I0203 10:07:33.157016 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/frr/0.log" Feb 03 10:07:33 crc kubenswrapper[4756]: I0203 10:07:33.264704 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-44db6_98d5a2ec-16f8-4654-b5ba-0c32b08d0fda/speaker/0.log" Feb 03 10:07:33 crc kubenswrapper[4756]: I0203 10:07:33.669840 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn9kj" event={"ID":"06a32867-0d73-4505-a0df-cf49e30369fa","Type":"ContainerStarted","Data":"68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e"} Feb 03 10:07:33 crc kubenswrapper[4756]: I0203 10:07:33.697969 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pn9kj" podStartSLOduration=2.156093424 podStartE2EDuration="4.697946647s" podCreationTimestamp="2026-02-03 10:07:29 +0000 UTC" firstStartedPulling="2026-02-03 10:07:30.609608565 +0000 UTC m=+3441.760075940" lastFinishedPulling="2026-02-03 10:07:33.151461788 +0000 UTC m=+3444.301929163" observedRunningTime="2026-02-03 10:07:33.688866603 +0000 UTC m=+3444.839333978" watchObservedRunningTime="2026-02-03 10:07:33.697946647 +0000 UTC m=+3444.848414022" Feb 03 10:07:39 crc kubenswrapper[4756]: I0203 10:07:39.453141 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:39 crc kubenswrapper[4756]: I0203 10:07:39.454940 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:39 crc kubenswrapper[4756]: I0203 10:07:39.501901 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:39 crc kubenswrapper[4756]: I0203 10:07:39.783284 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:39 crc kubenswrapper[4756]: I0203 10:07:39.832530 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pn9kj"] Feb 03 10:07:41 crc kubenswrapper[4756]: I0203 10:07:41.751125 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pn9kj" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" containerName="registry-server" containerID="cri-o://68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e" gracePeriod=2 Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.305813 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.380430 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-catalog-content\") pod \"06a32867-0d73-4505-a0df-cf49e30369fa\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.380607 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-utilities\") pod \"06a32867-0d73-4505-a0df-cf49e30369fa\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.380812 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m56ql\" (UniqueName: \"kubernetes.io/projected/06a32867-0d73-4505-a0df-cf49e30369fa-kube-api-access-m56ql\") pod \"06a32867-0d73-4505-a0df-cf49e30369fa\" (UID: \"06a32867-0d73-4505-a0df-cf49e30369fa\") " Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.382657 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-utilities" (OuterVolumeSpecName: "utilities") pod "06a32867-0d73-4505-a0df-cf49e30369fa" (UID: "06a32867-0d73-4505-a0df-cf49e30369fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.398057 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06a32867-0d73-4505-a0df-cf49e30369fa-kube-api-access-m56ql" (OuterVolumeSpecName: "kube-api-access-m56ql") pod "06a32867-0d73-4505-a0df-cf49e30369fa" (UID: "06a32867-0d73-4505-a0df-cf49e30369fa"). InnerVolumeSpecName "kube-api-access-m56ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.483806 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m56ql\" (UniqueName: \"kubernetes.io/projected/06a32867-0d73-4505-a0df-cf49e30369fa-kube-api-access-m56ql\") on node \"crc\" DevicePath \"\"" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.483872 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.764097 4756 generic.go:334] "Generic (PLEG): container finished" podID="06a32867-0d73-4505-a0df-cf49e30369fa" containerID="68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e" exitCode=0 Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.764167 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pn9kj" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.764201 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn9kj" event={"ID":"06a32867-0d73-4505-a0df-cf49e30369fa","Type":"ContainerDied","Data":"68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e"} Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.764662 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn9kj" event={"ID":"06a32867-0d73-4505-a0df-cf49e30369fa","Type":"ContainerDied","Data":"c4bb47f2999ed9e30474a5a03274b3c67f6f8710606be49f6a02090e61719d03"} Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.764694 4756 scope.go:117] "RemoveContainer" containerID="68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.789641 4756 scope.go:117] "RemoveContainer" containerID="e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.802936 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06a32867-0d73-4505-a0df-cf49e30369fa" (UID: "06a32867-0d73-4505-a0df-cf49e30369fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.810435 4756 scope.go:117] "RemoveContainer" containerID="69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.854236 4756 scope.go:117] "RemoveContainer" containerID="68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e" Feb 03 10:07:42 crc kubenswrapper[4756]: E0203 10:07:42.854717 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e\": container with ID starting with 68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e not found: ID does not exist" containerID="68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.854841 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e"} err="failed to get container status \"68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e\": rpc error: code = NotFound desc = could not find container \"68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e\": container with ID starting with 68052edaf271c3ebe9e0c661de12340563d5134c8f260a8c777634c613122a1e not found: ID does not exist" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.854974 4756 scope.go:117] "RemoveContainer" containerID="e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72" Feb 03 10:07:42 crc kubenswrapper[4756]: E0203 10:07:42.855509 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72\": container with ID starting with e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72 not found: ID does not exist" containerID="e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.855650 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72"} err="failed to get container status \"e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72\": rpc error: code = NotFound desc = could not find container \"e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72\": container with ID starting with e432633237b9f42aecf24ce35928928270767aa2565a9812c59669d2e5717c72 not found: ID does not exist" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.855756 4756 scope.go:117] "RemoveContainer" containerID="69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a" Feb 03 10:07:42 crc kubenswrapper[4756]: E0203 10:07:42.856112 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a\": container with ID starting with 69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a not found: ID does not exist" containerID="69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.856292 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a"} err="failed to get container status \"69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a\": rpc error: code = NotFound desc = could not find container \"69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a\": container with ID starting with 69bec70e7562f3daa2e848cbffb6b935cb8fd8defcfe347e6113c835e8d93e9a not found: ID does not exist" Feb 03 10:07:42 crc kubenswrapper[4756]: I0203 10:07:42.893263 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06a32867-0d73-4505-a0df-cf49e30369fa-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 10:07:43 crc kubenswrapper[4756]: I0203 10:07:43.098986 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pn9kj"] Feb 03 10:07:43 crc kubenswrapper[4756]: I0203 10:07:43.107185 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pn9kj"] Feb 03 10:07:43 crc kubenswrapper[4756]: I0203 10:07:43.626040 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" path="/var/lib/kubelet/pods/06a32867-0d73-4505-a0df-cf49e30369fa/volumes" Feb 03 10:07:46 crc kubenswrapper[4756]: I0203 10:07:46.405792 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/util/0.log" Feb 03 10:07:46 crc kubenswrapper[4756]: I0203 10:07:46.604878 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/util/0.log" Feb 03 10:07:46 crc kubenswrapper[4756]: I0203 10:07:46.605740 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/pull/0.log" Feb 03 10:07:46 crc kubenswrapper[4756]: I0203 10:07:46.639404 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/pull/0.log" Feb 03 10:07:46 crc kubenswrapper[4756]: I0203 10:07:46.806613 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/util/0.log" Feb 03 10:07:46 crc kubenswrapper[4756]: I0203 10:07:46.849614 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/pull/0.log" Feb 03 10:07:46 crc kubenswrapper[4756]: I0203 10:07:46.855767 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/extract/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.007108 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/util/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.150777 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/util/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.203071 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/pull/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.207101 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/pull/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.381367 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/pull/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.387126 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/util/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.455075 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/extract/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.558332 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-utilities/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.929030 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-content/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.934190 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-content/0.log" Feb 03 10:07:47 crc kubenswrapper[4756]: I0203 10:07:47.945110 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-utilities/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.172943 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-utilities/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.188517 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-content/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.465406 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-utilities/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.602411 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-utilities/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.634275 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-content/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.638103 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/registry-server/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.727435 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-content/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.848028 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-utilities/0.log" Feb 03 10:07:48 crc kubenswrapper[4756]: I0203 10:07:48.867867 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-content/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.059436 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xztlg_29956f45-fbd4-4630-82fe-8253dc497ce6/marketplace-operator/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.180247 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-utilities/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.429245 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/registry-server/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.521043 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-content/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.536812 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-content/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.556476 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-utilities/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.660502 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-utilities/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.733355 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-content/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.878339 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/registry-server/0.log" Feb 03 10:07:49 crc kubenswrapper[4756]: I0203 10:07:49.894959 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-utilities/0.log" Feb 03 10:07:50 crc kubenswrapper[4756]: I0203 10:07:50.099760 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-content/0.log" Feb 03 10:07:50 crc kubenswrapper[4756]: I0203 10:07:50.117110 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-utilities/0.log" Feb 03 10:07:50 crc kubenswrapper[4756]: I0203 10:07:50.130946 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-content/0.log" Feb 03 10:07:50 crc kubenswrapper[4756]: I0203 10:07:50.294808 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-utilities/0.log" Feb 03 10:07:50 crc kubenswrapper[4756]: I0203 10:07:50.315087 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-content/0.log" Feb 03 10:07:50 crc kubenswrapper[4756]: I0203 10:07:50.849883 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/registry-server/0.log" Feb 03 10:08:43 crc kubenswrapper[4756]: I0203 10:08:43.566480 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:08:43 crc kubenswrapper[4756]: I0203 10:08:43.567009 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:09:13 crc kubenswrapper[4756]: I0203 10:09:13.566426 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:09:13 crc kubenswrapper[4756]: I0203 10:09:13.566987 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:09:37 crc kubenswrapper[4756]: I0203 10:09:37.831135 4756 generic.go:334] "Generic (PLEG): container finished" podID="5a2ae189-29a7-41db-9916-910149f8b124" containerID="6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d" exitCode=0 Feb 03 10:09:37 crc kubenswrapper[4756]: I0203 10:09:37.831620 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5jjl/must-gather-ssccp" event={"ID":"5a2ae189-29a7-41db-9916-910149f8b124","Type":"ContainerDied","Data":"6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d"} Feb 03 10:09:37 crc kubenswrapper[4756]: I0203 10:09:37.832204 4756 scope.go:117] "RemoveContainer" containerID="6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d" Feb 03 10:09:38 crc kubenswrapper[4756]: I0203 10:09:38.326096 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h5jjl_must-gather-ssccp_5a2ae189-29a7-41db-9916-910149f8b124/gather/0.log" Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.566037 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.566638 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.566692 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.567373 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.567422 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" gracePeriod=600 Feb 03 10:09:43 crc kubenswrapper[4756]: E0203 10:09:43.717559 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.883772 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" exitCode=0 Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.883818 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3"} Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.883853 4756 scope.go:117] "RemoveContainer" containerID="e93ac21e9eba4fd7bad1c059d84ee34641b9f4a108f333d3d5f53ecefa6ab116" Feb 03 10:09:43 crc kubenswrapper[4756]: I0203 10:09:43.884594 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:09:43 crc kubenswrapper[4756]: E0203 10:09:43.884884 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.145196 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h5jjl/must-gather-ssccp"] Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.145907 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-h5jjl/must-gather-ssccp" podUID="5a2ae189-29a7-41db-9916-910149f8b124" containerName="copy" containerID="cri-o://322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8" gracePeriod=2 Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.158766 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h5jjl/must-gather-ssccp"] Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.569944 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h5jjl_must-gather-ssccp_5a2ae189-29a7-41db-9916-910149f8b124/copy/0.log" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.570679 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.726502 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnrb4\" (UniqueName: \"kubernetes.io/projected/5a2ae189-29a7-41db-9916-910149f8b124-kube-api-access-hnrb4\") pod \"5a2ae189-29a7-41db-9916-910149f8b124\" (UID: \"5a2ae189-29a7-41db-9916-910149f8b124\") " Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.727201 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5a2ae189-29a7-41db-9916-910149f8b124-must-gather-output\") pod \"5a2ae189-29a7-41db-9916-910149f8b124\" (UID: \"5a2ae189-29a7-41db-9916-910149f8b124\") " Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.733488 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a2ae189-29a7-41db-9916-910149f8b124-kube-api-access-hnrb4" (OuterVolumeSpecName: "kube-api-access-hnrb4") pod "5a2ae189-29a7-41db-9916-910149f8b124" (UID: "5a2ae189-29a7-41db-9916-910149f8b124"). InnerVolumeSpecName "kube-api-access-hnrb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.828949 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnrb4\" (UniqueName: \"kubernetes.io/projected/5a2ae189-29a7-41db-9916-910149f8b124-kube-api-access-hnrb4\") on node \"crc\" DevicePath \"\"" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.872675 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a2ae189-29a7-41db-9916-910149f8b124-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5a2ae189-29a7-41db-9916-910149f8b124" (UID: "5a2ae189-29a7-41db-9916-910149f8b124"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.920253 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h5jjl_must-gather-ssccp_5a2ae189-29a7-41db-9916-910149f8b124/copy/0.log" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.920816 4756 generic.go:334] "Generic (PLEG): container finished" podID="5a2ae189-29a7-41db-9916-910149f8b124" containerID="322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8" exitCode=143 Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.920933 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5jjl/must-gather-ssccp" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.920933 4756 scope.go:117] "RemoveContainer" containerID="322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.930614 4756 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5a2ae189-29a7-41db-9916-910149f8b124-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 03 10:09:46 crc kubenswrapper[4756]: I0203 10:09:46.950856 4756 scope.go:117] "RemoveContainer" containerID="6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d" Feb 03 10:09:47 crc kubenswrapper[4756]: I0203 10:09:47.048091 4756 scope.go:117] "RemoveContainer" containerID="322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8" Feb 03 10:09:47 crc kubenswrapper[4756]: E0203 10:09:47.048617 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8\": container with ID starting with 322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8 not found: ID does not exist" containerID="322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8" Feb 03 10:09:47 crc kubenswrapper[4756]: I0203 10:09:47.048725 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8"} err="failed to get container status \"322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8\": rpc error: code = NotFound desc = could not find container \"322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8\": container with ID starting with 322c342670f63c3ac66fcfa952ae337bee75eaf2f7368462b4c4e56d036ea7f8 not found: ID does not exist" Feb 03 10:09:47 crc kubenswrapper[4756]: I0203 10:09:47.048845 4756 scope.go:117] "RemoveContainer" containerID="6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d" Feb 03 10:09:47 crc kubenswrapper[4756]: E0203 10:09:47.049223 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d\": container with ID starting with 6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d not found: ID does not exist" containerID="6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d" Feb 03 10:09:47 crc kubenswrapper[4756]: I0203 10:09:47.049317 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d"} err="failed to get container status \"6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d\": rpc error: code = NotFound desc = could not find container \"6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d\": container with ID starting with 6545d1fd0b506471b148e853b879bc4b68bd91624a69d1f025ff6afc2482fb5d not found: ID does not exist" Feb 03 10:09:47 crc kubenswrapper[4756]: I0203 10:09:47.634606 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a2ae189-29a7-41db-9916-910149f8b124" path="/var/lib/kubelet/pods/5a2ae189-29a7-41db-9916-910149f8b124/volumes" Feb 03 10:09:56 crc kubenswrapper[4756]: I0203 10:09:56.614577 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:09:56 crc kubenswrapper[4756]: E0203 10:09:56.615435 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:10:10 crc kubenswrapper[4756]: I0203 10:10:10.614232 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:10:10 crc kubenswrapper[4756]: E0203 10:10:10.614936 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:10:21 crc kubenswrapper[4756]: I0203 10:10:21.615736 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:10:21 crc kubenswrapper[4756]: E0203 10:10:21.617080 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:10:32 crc kubenswrapper[4756]: I0203 10:10:32.087160 4756 scope.go:117] "RemoveContainer" containerID="b64101f92ad09a0506e4e3e26cbfac81e9be9e0973570d654d1e4abffdfae6b9" Feb 03 10:10:36 crc kubenswrapper[4756]: I0203 10:10:36.615018 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:10:36 crc kubenswrapper[4756]: E0203 10:10:36.615830 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.407873 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kp4vw"] Feb 03 10:10:42 crc kubenswrapper[4756]: E0203 10:10:42.408866 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" containerName="extract-utilities" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.408885 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" containerName="extract-utilities" Feb 03 10:10:42 crc kubenswrapper[4756]: E0203 10:10:42.408906 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" containerName="extract-content" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.408914 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" containerName="extract-content" Feb 03 10:10:42 crc kubenswrapper[4756]: E0203 10:10:42.408923 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" containerName="registry-server" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.408930 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" containerName="registry-server" Feb 03 10:10:42 crc kubenswrapper[4756]: E0203 10:10:42.408946 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2ae189-29a7-41db-9916-910149f8b124" containerName="gather" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.408953 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2ae189-29a7-41db-9916-910149f8b124" containerName="gather" Feb 03 10:10:42 crc kubenswrapper[4756]: E0203 10:10:42.408963 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2ae189-29a7-41db-9916-910149f8b124" containerName="copy" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.408970 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2ae189-29a7-41db-9916-910149f8b124" containerName="copy" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.409194 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="06a32867-0d73-4505-a0df-cf49e30369fa" containerName="registry-server" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.409210 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a2ae189-29a7-41db-9916-910149f8b124" containerName="gather" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.409223 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a2ae189-29a7-41db-9916-910149f8b124" containerName="copy" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.410881 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.419404 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp4vw"] Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.521623 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv2b5\" (UniqueName: \"kubernetes.io/projected/6dc82eea-5055-4b96-8eee-258dbef1e007-kube-api-access-dv2b5\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.521947 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-utilities\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.522107 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-catalog-content\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.623745 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv2b5\" (UniqueName: \"kubernetes.io/projected/6dc82eea-5055-4b96-8eee-258dbef1e007-kube-api-access-dv2b5\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.623849 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-utilities\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.623890 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-catalog-content\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.624526 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-catalog-content\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.624656 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-utilities\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.648126 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv2b5\" (UniqueName: \"kubernetes.io/projected/6dc82eea-5055-4b96-8eee-258dbef1e007-kube-api-access-dv2b5\") pod \"redhat-marketplace-kp4vw\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:42 crc kubenswrapper[4756]: I0203 10:10:42.732206 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:43 crc kubenswrapper[4756]: I0203 10:10:43.194162 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp4vw"] Feb 03 10:10:43 crc kubenswrapper[4756]: I0203 10:10:43.439119 4756 generic.go:334] "Generic (PLEG): container finished" podID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerID="c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384" exitCode=0 Feb 03 10:10:43 crc kubenswrapper[4756]: I0203 10:10:43.439172 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp4vw" event={"ID":"6dc82eea-5055-4b96-8eee-258dbef1e007","Type":"ContainerDied","Data":"c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384"} Feb 03 10:10:43 crc kubenswrapper[4756]: I0203 10:10:43.439205 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp4vw" event={"ID":"6dc82eea-5055-4b96-8eee-258dbef1e007","Type":"ContainerStarted","Data":"13a94a454c031018c3473e42cec5824d58e1e4ef33e87d32d32b2a204e7f2bb6"} Feb 03 10:10:43 crc kubenswrapper[4756]: I0203 10:10:43.441133 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 10:10:44 crc kubenswrapper[4756]: I0203 10:10:44.449246 4756 generic.go:334] "Generic (PLEG): container finished" podID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerID="b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785" exitCode=0 Feb 03 10:10:44 crc kubenswrapper[4756]: I0203 10:10:44.449320 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp4vw" event={"ID":"6dc82eea-5055-4b96-8eee-258dbef1e007","Type":"ContainerDied","Data":"b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785"} Feb 03 10:10:45 crc kubenswrapper[4756]: I0203 10:10:45.460937 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp4vw" event={"ID":"6dc82eea-5055-4b96-8eee-258dbef1e007","Type":"ContainerStarted","Data":"382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22"} Feb 03 10:10:45 crc kubenswrapper[4756]: I0203 10:10:45.482270 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kp4vw" podStartSLOduration=1.9637411120000001 podStartE2EDuration="3.482250872s" podCreationTimestamp="2026-02-03 10:10:42 +0000 UTC" firstStartedPulling="2026-02-03 10:10:43.440907752 +0000 UTC m=+3634.591375127" lastFinishedPulling="2026-02-03 10:10:44.959417512 +0000 UTC m=+3636.109884887" observedRunningTime="2026-02-03 10:10:45.4764211 +0000 UTC m=+3636.626888505" watchObservedRunningTime="2026-02-03 10:10:45.482250872 +0000 UTC m=+3636.632718247" Feb 03 10:10:48 crc kubenswrapper[4756]: I0203 10:10:48.615090 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:10:48 crc kubenswrapper[4756]: E0203 10:10:48.615731 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:10:52 crc kubenswrapper[4756]: I0203 10:10:52.733269 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:52 crc kubenswrapper[4756]: I0203 10:10:52.734137 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:52 crc kubenswrapper[4756]: I0203 10:10:52.776079 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:53 crc kubenswrapper[4756]: I0203 10:10:53.563813 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:53 crc kubenswrapper[4756]: I0203 10:10:53.612733 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp4vw"] Feb 03 10:10:55 crc kubenswrapper[4756]: I0203 10:10:55.536529 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kp4vw" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerName="registry-server" containerID="cri-o://382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22" gracePeriod=2 Feb 03 10:10:55 crc kubenswrapper[4756]: I0203 10:10:55.982664 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.004084 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv2b5\" (UniqueName: \"kubernetes.io/projected/6dc82eea-5055-4b96-8eee-258dbef1e007-kube-api-access-dv2b5\") pod \"6dc82eea-5055-4b96-8eee-258dbef1e007\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.004163 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-catalog-content\") pod \"6dc82eea-5055-4b96-8eee-258dbef1e007\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.004207 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-utilities\") pod \"6dc82eea-5055-4b96-8eee-258dbef1e007\" (UID: \"6dc82eea-5055-4b96-8eee-258dbef1e007\") " Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.005717 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-utilities" (OuterVolumeSpecName: "utilities") pod "6dc82eea-5055-4b96-8eee-258dbef1e007" (UID: "6dc82eea-5055-4b96-8eee-258dbef1e007"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.034228 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dc82eea-5055-4b96-8eee-258dbef1e007-kube-api-access-dv2b5" (OuterVolumeSpecName: "kube-api-access-dv2b5") pod "6dc82eea-5055-4b96-8eee-258dbef1e007" (UID: "6dc82eea-5055-4b96-8eee-258dbef1e007"). InnerVolumeSpecName "kube-api-access-dv2b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.064370 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6dc82eea-5055-4b96-8eee-258dbef1e007" (UID: "6dc82eea-5055-4b96-8eee-258dbef1e007"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.105723 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv2b5\" (UniqueName: \"kubernetes.io/projected/6dc82eea-5055-4b96-8eee-258dbef1e007-kube-api-access-dv2b5\") on node \"crc\" DevicePath \"\"" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.106054 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.106073 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6dc82eea-5055-4b96-8eee-258dbef1e007-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.550715 4756 generic.go:334] "Generic (PLEG): container finished" podID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerID="382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22" exitCode=0 Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.550776 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp4vw" event={"ID":"6dc82eea-5055-4b96-8eee-258dbef1e007","Type":"ContainerDied","Data":"382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22"} Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.550799 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kp4vw" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.550824 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kp4vw" event={"ID":"6dc82eea-5055-4b96-8eee-258dbef1e007","Type":"ContainerDied","Data":"13a94a454c031018c3473e42cec5824d58e1e4ef33e87d32d32b2a204e7f2bb6"} Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.550851 4756 scope.go:117] "RemoveContainer" containerID="382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.576037 4756 scope.go:117] "RemoveContainer" containerID="b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.597073 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp4vw"] Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.610433 4756 scope.go:117] "RemoveContainer" containerID="c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.615141 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kp4vw"] Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.655465 4756 scope.go:117] "RemoveContainer" containerID="382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22" Feb 03 10:10:56 crc kubenswrapper[4756]: E0203 10:10:56.655966 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22\": container with ID starting with 382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22 not found: ID does not exist" containerID="382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.656005 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22"} err="failed to get container status \"382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22\": rpc error: code = NotFound desc = could not find container \"382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22\": container with ID starting with 382f52548834d1632f25c77d909c5816573f8c376c594c90cd90fff4a8a2ac22 not found: ID does not exist" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.656025 4756 scope.go:117] "RemoveContainer" containerID="b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785" Feb 03 10:10:56 crc kubenswrapper[4756]: E0203 10:10:56.656326 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785\": container with ID starting with b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785 not found: ID does not exist" containerID="b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.656354 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785"} err="failed to get container status \"b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785\": rpc error: code = NotFound desc = could not find container \"b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785\": container with ID starting with b92a20e33bcdf64bfda05149ce3ab54777e0e31ea7df3c198df1c2e31a19e785 not found: ID does not exist" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.656371 4756 scope.go:117] "RemoveContainer" containerID="c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384" Feb 03 10:10:56 crc kubenswrapper[4756]: E0203 10:10:56.656712 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384\": container with ID starting with c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384 not found: ID does not exist" containerID="c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384" Feb 03 10:10:56 crc kubenswrapper[4756]: I0203 10:10:56.656736 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384"} err="failed to get container status \"c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384\": rpc error: code = NotFound desc = could not find container \"c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384\": container with ID starting with c61431cb22b465ea40b9a2b134dca2e6dd365d8ba5bea601a729d05d0b414384 not found: ID does not exist" Feb 03 10:10:57 crc kubenswrapper[4756]: I0203 10:10:57.624918 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" path="/var/lib/kubelet/pods/6dc82eea-5055-4b96-8eee-258dbef1e007/volumes" Feb 03 10:11:01 crc kubenswrapper[4756]: I0203 10:11:01.613883 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:11:01 crc kubenswrapper[4756]: E0203 10:11:01.614715 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:11:13 crc kubenswrapper[4756]: I0203 10:11:13.614559 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:11:13 crc kubenswrapper[4756]: E0203 10:11:13.616781 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:11:28 crc kubenswrapper[4756]: I0203 10:11:28.614373 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:11:28 crc kubenswrapper[4756]: E0203 10:11:28.615253 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:11:43 crc kubenswrapper[4756]: I0203 10:11:43.614860 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:11:43 crc kubenswrapper[4756]: E0203 10:11:43.615712 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:11:54 crc kubenswrapper[4756]: I0203 10:11:54.614766 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:11:54 crc kubenswrapper[4756]: E0203 10:11:54.616580 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.370426 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9x4db"] Feb 03 10:12:00 crc kubenswrapper[4756]: E0203 10:12:00.371415 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerName="extract-content" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.371433 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerName="extract-content" Feb 03 10:12:00 crc kubenswrapper[4756]: E0203 10:12:00.371475 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerName="extract-utilities" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.371484 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerName="extract-utilities" Feb 03 10:12:00 crc kubenswrapper[4756]: E0203 10:12:00.371505 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerName="registry-server" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.371515 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerName="registry-server" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.371804 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dc82eea-5055-4b96-8eee-258dbef1e007" containerName="registry-server" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.373336 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.397589 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9x4db"] Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.511301 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jn2d\" (UniqueName: \"kubernetes.io/projected/c534f146-375f-4f3b-b718-83bbb3fe3bcb-kube-api-access-5jn2d\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.511649 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-utilities\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.511850 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-catalog-content\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.614140 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jn2d\" (UniqueName: \"kubernetes.io/projected/c534f146-375f-4f3b-b718-83bbb3fe3bcb-kube-api-access-5jn2d\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.614231 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-utilities\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.614266 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-catalog-content\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.614866 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-catalog-content\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.614907 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-utilities\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.633544 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jn2d\" (UniqueName: \"kubernetes.io/projected/c534f146-375f-4f3b-b718-83bbb3fe3bcb-kube-api-access-5jn2d\") pod \"certified-operators-9x4db\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:00 crc kubenswrapper[4756]: I0203 10:12:00.704020 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:01 crc kubenswrapper[4756]: I0203 10:12:01.209016 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9x4db"] Feb 03 10:12:02 crc kubenswrapper[4756]: I0203 10:12:02.088587 4756 generic.go:334] "Generic (PLEG): container finished" podID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerID="f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea" exitCode=0 Feb 03 10:12:02 crc kubenswrapper[4756]: I0203 10:12:02.088686 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x4db" event={"ID":"c534f146-375f-4f3b-b718-83bbb3fe3bcb","Type":"ContainerDied","Data":"f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea"} Feb 03 10:12:02 crc kubenswrapper[4756]: I0203 10:12:02.088910 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x4db" event={"ID":"c534f146-375f-4f3b-b718-83bbb3fe3bcb","Type":"ContainerStarted","Data":"f9237e7ac6761c4e4897247745d2cd8d21263f659a5aab367d9b22c15d9cc8dc"} Feb 03 10:12:04 crc kubenswrapper[4756]: I0203 10:12:04.109115 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x4db" event={"ID":"c534f146-375f-4f3b-b718-83bbb3fe3bcb","Type":"ContainerStarted","Data":"fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0"} Feb 03 10:12:05 crc kubenswrapper[4756]: I0203 10:12:05.117973 4756 generic.go:334] "Generic (PLEG): container finished" podID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerID="fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0" exitCode=0 Feb 03 10:12:05 crc kubenswrapper[4756]: I0203 10:12:05.118038 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x4db" event={"ID":"c534f146-375f-4f3b-b718-83bbb3fe3bcb","Type":"ContainerDied","Data":"fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0"} Feb 03 10:12:05 crc kubenswrapper[4756]: I0203 10:12:05.986257 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:12:05 crc kubenswrapper[4756]: E0203 10:12:05.987480 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:12:07 crc kubenswrapper[4756]: I0203 10:12:07.141161 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x4db" event={"ID":"c534f146-375f-4f3b-b718-83bbb3fe3bcb","Type":"ContainerStarted","Data":"8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50"} Feb 03 10:12:07 crc kubenswrapper[4756]: I0203 10:12:07.165005 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9x4db" podStartSLOduration=3.232719162 podStartE2EDuration="7.164982117s" podCreationTimestamp="2026-02-03 10:12:00 +0000 UTC" firstStartedPulling="2026-02-03 10:12:02.091111897 +0000 UTC m=+3713.241579272" lastFinishedPulling="2026-02-03 10:12:06.023374852 +0000 UTC m=+3717.173842227" observedRunningTime="2026-02-03 10:12:07.158055531 +0000 UTC m=+3718.308522906" watchObservedRunningTime="2026-02-03 10:12:07.164982117 +0000 UTC m=+3718.315449512" Feb 03 10:12:10 crc kubenswrapper[4756]: I0203 10:12:10.704367 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:10 crc kubenswrapper[4756]: I0203 10:12:10.704691 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:10 crc kubenswrapper[4756]: I0203 10:12:10.746820 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:11 crc kubenswrapper[4756]: I0203 10:12:11.234025 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:11 crc kubenswrapper[4756]: I0203 10:12:11.283307 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9x4db"] Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.199677 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9x4db" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerName="registry-server" containerID="cri-o://8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50" gracePeriod=2 Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.630499 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.774334 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-utilities\") pod \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.774505 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-catalog-content\") pod \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.774531 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jn2d\" (UniqueName: \"kubernetes.io/projected/c534f146-375f-4f3b-b718-83bbb3fe3bcb-kube-api-access-5jn2d\") pod \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\" (UID: \"c534f146-375f-4f3b-b718-83bbb3fe3bcb\") " Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.775240 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-utilities" (OuterVolumeSpecName: "utilities") pod "c534f146-375f-4f3b-b718-83bbb3fe3bcb" (UID: "c534f146-375f-4f3b-b718-83bbb3fe3bcb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.782264 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c534f146-375f-4f3b-b718-83bbb3fe3bcb-kube-api-access-5jn2d" (OuterVolumeSpecName: "kube-api-access-5jn2d") pod "c534f146-375f-4f3b-b718-83bbb3fe3bcb" (UID: "c534f146-375f-4f3b-b718-83bbb3fe3bcb"). InnerVolumeSpecName "kube-api-access-5jn2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.842059 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c534f146-375f-4f3b-b718-83bbb3fe3bcb" (UID: "c534f146-375f-4f3b-b718-83bbb3fe3bcb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.876429 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.876498 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c534f146-375f-4f3b-b718-83bbb3fe3bcb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 10:12:13 crc kubenswrapper[4756]: I0203 10:12:13.876511 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jn2d\" (UniqueName: \"kubernetes.io/projected/c534f146-375f-4f3b-b718-83bbb3fe3bcb-kube-api-access-5jn2d\") on node \"crc\" DevicePath \"\"" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.210169 4756 generic.go:334] "Generic (PLEG): container finished" podID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerID="8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50" exitCode=0 Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.210243 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x4db" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.210269 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x4db" event={"ID":"c534f146-375f-4f3b-b718-83bbb3fe3bcb","Type":"ContainerDied","Data":"8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50"} Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.210646 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x4db" event={"ID":"c534f146-375f-4f3b-b718-83bbb3fe3bcb","Type":"ContainerDied","Data":"f9237e7ac6761c4e4897247745d2cd8d21263f659a5aab367d9b22c15d9cc8dc"} Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.210665 4756 scope.go:117] "RemoveContainer" containerID="8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.229504 4756 scope.go:117] "RemoveContainer" containerID="fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.245004 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9x4db"] Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.257230 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9x4db"] Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.261790 4756 scope.go:117] "RemoveContainer" containerID="f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.299606 4756 scope.go:117] "RemoveContainer" containerID="8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50" Feb 03 10:12:14 crc kubenswrapper[4756]: E0203 10:12:14.300109 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50\": container with ID starting with 8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50 not found: ID does not exist" containerID="8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.300165 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50"} err="failed to get container status \"8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50\": rpc error: code = NotFound desc = could not find container \"8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50\": container with ID starting with 8a95ce47218e7f535118137105fb97d8c42f3dc5c8abcc31cadd3615f281de50 not found: ID does not exist" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.300190 4756 scope.go:117] "RemoveContainer" containerID="fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0" Feb 03 10:12:14 crc kubenswrapper[4756]: E0203 10:12:14.300615 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0\": container with ID starting with fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0 not found: ID does not exist" containerID="fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.300644 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0"} err="failed to get container status \"fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0\": rpc error: code = NotFound desc = could not find container \"fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0\": container with ID starting with fcf9db466bfbfd74176be7e99230a9c7bc0c44593c7771426a048adcb6c0f5d0 not found: ID does not exist" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.300663 4756 scope.go:117] "RemoveContainer" containerID="f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea" Feb 03 10:12:14 crc kubenswrapper[4756]: E0203 10:12:14.300996 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea\": container with ID starting with f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea not found: ID does not exist" containerID="f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea" Feb 03 10:12:14 crc kubenswrapper[4756]: I0203 10:12:14.301023 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea"} err="failed to get container status \"f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea\": rpc error: code = NotFound desc = could not find container \"f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea\": container with ID starting with f2c1e4bb0eab54dea56bbf94523b3f5e51ff251a0978f1face2846436c69faea not found: ID does not exist" Feb 03 10:12:15 crc kubenswrapper[4756]: I0203 10:12:15.624235 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" path="/var/lib/kubelet/pods/c534f146-375f-4f3b-b718-83bbb3fe3bcb/volumes" Feb 03 10:12:18 crc kubenswrapper[4756]: I0203 10:12:18.615013 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:12:18 crc kubenswrapper[4756]: E0203 10:12:18.615601 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:12:32 crc kubenswrapper[4756]: I0203 10:12:32.614025 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:12:32 crc kubenswrapper[4756]: E0203 10:12:32.614752 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:12:47 crc kubenswrapper[4756]: I0203 10:12:47.614625 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:12:47 crc kubenswrapper[4756]: E0203 10:12:47.615544 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.147877 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gc6f/must-gather-c58nj"] Feb 03 10:12:57 crc kubenswrapper[4756]: E0203 10:12:57.148756 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerName="extract-content" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.148768 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerName="extract-content" Feb 03 10:12:57 crc kubenswrapper[4756]: E0203 10:12:57.148799 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerName="registry-server" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.148806 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerName="registry-server" Feb 03 10:12:57 crc kubenswrapper[4756]: E0203 10:12:57.148818 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerName="extract-utilities" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.148824 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerName="extract-utilities" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.149068 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="c534f146-375f-4f3b-b718-83bbb3fe3bcb" containerName="registry-server" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.150057 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.153556 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4gc6f"/"kube-root-ca.crt" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.153754 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4gc6f"/"openshift-service-ca.crt" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.170292 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4gc6f/must-gather-c58nj"] Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.304259 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4r77\" (UniqueName: \"kubernetes.io/projected/c56ae5f4-06ed-41ff-84e2-093504936086-kube-api-access-n4r77\") pod \"must-gather-c58nj\" (UID: \"c56ae5f4-06ed-41ff-84e2-093504936086\") " pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.304326 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c56ae5f4-06ed-41ff-84e2-093504936086-must-gather-output\") pod \"must-gather-c58nj\" (UID: \"c56ae5f4-06ed-41ff-84e2-093504936086\") " pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.405799 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c56ae5f4-06ed-41ff-84e2-093504936086-must-gather-output\") pod \"must-gather-c58nj\" (UID: \"c56ae5f4-06ed-41ff-84e2-093504936086\") " pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.405962 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4r77\" (UniqueName: \"kubernetes.io/projected/c56ae5f4-06ed-41ff-84e2-093504936086-kube-api-access-n4r77\") pod \"must-gather-c58nj\" (UID: \"c56ae5f4-06ed-41ff-84e2-093504936086\") " pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.406371 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c56ae5f4-06ed-41ff-84e2-093504936086-must-gather-output\") pod \"must-gather-c58nj\" (UID: \"c56ae5f4-06ed-41ff-84e2-093504936086\") " pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.425137 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4r77\" (UniqueName: \"kubernetes.io/projected/c56ae5f4-06ed-41ff-84e2-093504936086-kube-api-access-n4r77\") pod \"must-gather-c58nj\" (UID: \"c56ae5f4-06ed-41ff-84e2-093504936086\") " pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.470829 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:12:57 crc kubenswrapper[4756]: I0203 10:12:57.932123 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4gc6f/must-gather-c58nj"] Feb 03 10:12:58 crc kubenswrapper[4756]: I0203 10:12:58.580905 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/must-gather-c58nj" event={"ID":"c56ae5f4-06ed-41ff-84e2-093504936086","Type":"ContainerStarted","Data":"d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c"} Feb 03 10:12:58 crc kubenswrapper[4756]: I0203 10:12:58.581165 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/must-gather-c58nj" event={"ID":"c56ae5f4-06ed-41ff-84e2-093504936086","Type":"ContainerStarted","Data":"7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159"} Feb 03 10:12:58 crc kubenswrapper[4756]: I0203 10:12:58.581175 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/must-gather-c58nj" event={"ID":"c56ae5f4-06ed-41ff-84e2-093504936086","Type":"ContainerStarted","Data":"602255abbe8c2f803c4f84f6cde0814c08a47e412b08bbd6f7965eb1ca0f8edd"} Feb 03 10:12:58 crc kubenswrapper[4756]: I0203 10:12:58.603717 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4gc6f/must-gather-c58nj" podStartSLOduration=1.603693678 podStartE2EDuration="1.603693678s" podCreationTimestamp="2026-02-03 10:12:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 10:12:58.59991196 +0000 UTC m=+3769.750379335" watchObservedRunningTime="2026-02-03 10:12:58.603693678 +0000 UTC m=+3769.754161053" Feb 03 10:12:58 crc kubenswrapper[4756]: I0203 10:12:58.614549 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:12:58 crc kubenswrapper[4756]: E0203 10:12:58.614804 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:13:01 crc kubenswrapper[4756]: I0203 10:13:01.777366 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-vprsx"] Feb 03 10:13:01 crc kubenswrapper[4756]: I0203 10:13:01.778966 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:01 crc kubenswrapper[4756]: I0203 10:13:01.780991 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4gc6f"/"default-dockercfg-5d8m8" Feb 03 10:13:01 crc kubenswrapper[4756]: I0203 10:13:01.891353 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da0736cb-84e1-46ad-853a-b4ac5b25afaa-host\") pod \"crc-debug-vprsx\" (UID: \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\") " pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:01 crc kubenswrapper[4756]: I0203 10:13:01.891751 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s94gg\" (UniqueName: \"kubernetes.io/projected/da0736cb-84e1-46ad-853a-b4ac5b25afaa-kube-api-access-s94gg\") pod \"crc-debug-vprsx\" (UID: \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\") " pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:01 crc kubenswrapper[4756]: I0203 10:13:01.992873 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da0736cb-84e1-46ad-853a-b4ac5b25afaa-host\") pod \"crc-debug-vprsx\" (UID: \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\") " pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:01 crc kubenswrapper[4756]: I0203 10:13:01.992944 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s94gg\" (UniqueName: \"kubernetes.io/projected/da0736cb-84e1-46ad-853a-b4ac5b25afaa-kube-api-access-s94gg\") pod \"crc-debug-vprsx\" (UID: \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\") " pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:01 crc kubenswrapper[4756]: I0203 10:13:01.993039 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da0736cb-84e1-46ad-853a-b4ac5b25afaa-host\") pod \"crc-debug-vprsx\" (UID: \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\") " pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:02 crc kubenswrapper[4756]: I0203 10:13:02.015820 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s94gg\" (UniqueName: \"kubernetes.io/projected/da0736cb-84e1-46ad-853a-b4ac5b25afaa-kube-api-access-s94gg\") pod \"crc-debug-vprsx\" (UID: \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\") " pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:02 crc kubenswrapper[4756]: I0203 10:13:02.100962 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:02 crc kubenswrapper[4756]: W0203 10:13:02.127636 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda0736cb_84e1_46ad_853a_b4ac5b25afaa.slice/crio-0c00f5d71968a84212d4a7e2399c72deb9f224a6e3ebf771f221b089676a1508 WatchSource:0}: Error finding container 0c00f5d71968a84212d4a7e2399c72deb9f224a6e3ebf771f221b089676a1508: Status 404 returned error can't find the container with id 0c00f5d71968a84212d4a7e2399c72deb9f224a6e3ebf771f221b089676a1508 Feb 03 10:13:02 crc kubenswrapper[4756]: I0203 10:13:02.619498 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/crc-debug-vprsx" event={"ID":"da0736cb-84e1-46ad-853a-b4ac5b25afaa","Type":"ContainerStarted","Data":"e3155261dcaeda50af721ff22b566956e4c5e4cb8ff6cf268aa7b87954ef6aa2"} Feb 03 10:13:02 crc kubenswrapper[4756]: I0203 10:13:02.620002 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/crc-debug-vprsx" event={"ID":"da0736cb-84e1-46ad-853a-b4ac5b25afaa","Type":"ContainerStarted","Data":"0c00f5d71968a84212d4a7e2399c72deb9f224a6e3ebf771f221b089676a1508"} Feb 03 10:13:02 crc kubenswrapper[4756]: I0203 10:13:02.636800 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4gc6f/crc-debug-vprsx" podStartSLOduration=1.63678116 podStartE2EDuration="1.63678116s" podCreationTimestamp="2026-02-03 10:13:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 10:13:02.634924012 +0000 UTC m=+3773.785391387" watchObservedRunningTime="2026-02-03 10:13:02.63678116 +0000 UTC m=+3773.787248545" Feb 03 10:13:09 crc kubenswrapper[4756]: I0203 10:13:09.637920 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:13:09 crc kubenswrapper[4756]: E0203 10:13:09.639069 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:13:20 crc kubenswrapper[4756]: I0203 10:13:20.614366 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:13:20 crc kubenswrapper[4756]: E0203 10:13:20.615245 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:13:32 crc kubenswrapper[4756]: I0203 10:13:32.614580 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:13:32 crc kubenswrapper[4756]: E0203 10:13:32.615389 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:13:38 crc kubenswrapper[4756]: I0203 10:13:38.635876 4756 generic.go:334] "Generic (PLEG): container finished" podID="da0736cb-84e1-46ad-853a-b4ac5b25afaa" containerID="e3155261dcaeda50af721ff22b566956e4c5e4cb8ff6cf268aa7b87954ef6aa2" exitCode=0 Feb 03 10:13:38 crc kubenswrapper[4756]: I0203 10:13:38.635972 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/crc-debug-vprsx" event={"ID":"da0736cb-84e1-46ad-853a-b4ac5b25afaa","Type":"ContainerDied","Data":"e3155261dcaeda50af721ff22b566956e4c5e4cb8ff6cf268aa7b87954ef6aa2"} Feb 03 10:13:39 crc kubenswrapper[4756]: I0203 10:13:39.764406 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:39 crc kubenswrapper[4756]: I0203 10:13:39.794057 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-vprsx"] Feb 03 10:13:39 crc kubenswrapper[4756]: I0203 10:13:39.802088 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-vprsx"] Feb 03 10:13:39 crc kubenswrapper[4756]: I0203 10:13:39.931076 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da0736cb-84e1-46ad-853a-b4ac5b25afaa-host\") pod \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\" (UID: \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\") " Feb 03 10:13:39 crc kubenswrapper[4756]: I0203 10:13:39.931173 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da0736cb-84e1-46ad-853a-b4ac5b25afaa-host" (OuterVolumeSpecName: "host") pod "da0736cb-84e1-46ad-853a-b4ac5b25afaa" (UID: "da0736cb-84e1-46ad-853a-b4ac5b25afaa"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 10:13:39 crc kubenswrapper[4756]: I0203 10:13:39.932625 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s94gg\" (UniqueName: \"kubernetes.io/projected/da0736cb-84e1-46ad-853a-b4ac5b25afaa-kube-api-access-s94gg\") pod \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\" (UID: \"da0736cb-84e1-46ad-853a-b4ac5b25afaa\") " Feb 03 10:13:39 crc kubenswrapper[4756]: I0203 10:13:39.933573 4756 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da0736cb-84e1-46ad-853a-b4ac5b25afaa-host\") on node \"crc\" DevicePath \"\"" Feb 03 10:13:39 crc kubenswrapper[4756]: I0203 10:13:39.937654 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da0736cb-84e1-46ad-853a-b4ac5b25afaa-kube-api-access-s94gg" (OuterVolumeSpecName: "kube-api-access-s94gg") pod "da0736cb-84e1-46ad-853a-b4ac5b25afaa" (UID: "da0736cb-84e1-46ad-853a-b4ac5b25afaa"). InnerVolumeSpecName "kube-api-access-s94gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:13:40 crc kubenswrapper[4756]: I0203 10:13:40.035537 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s94gg\" (UniqueName: \"kubernetes.io/projected/da0736cb-84e1-46ad-853a-b4ac5b25afaa-kube-api-access-s94gg\") on node \"crc\" DevicePath \"\"" Feb 03 10:13:40 crc kubenswrapper[4756]: I0203 10:13:40.657167 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c00f5d71968a84212d4a7e2399c72deb9f224a6e3ebf771f221b089676a1508" Feb 03 10:13:40 crc kubenswrapper[4756]: I0203 10:13:40.657383 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-vprsx" Feb 03 10:13:40 crc kubenswrapper[4756]: I0203 10:13:40.988345 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-mjkv4"] Feb 03 10:13:40 crc kubenswrapper[4756]: E0203 10:13:40.988756 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da0736cb-84e1-46ad-853a-b4ac5b25afaa" containerName="container-00" Feb 03 10:13:40 crc kubenswrapper[4756]: I0203 10:13:40.988770 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="da0736cb-84e1-46ad-853a-b4ac5b25afaa" containerName="container-00" Feb 03 10:13:40 crc kubenswrapper[4756]: I0203 10:13:40.988954 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="da0736cb-84e1-46ad-853a-b4ac5b25afaa" containerName="container-00" Feb 03 10:13:40 crc kubenswrapper[4756]: I0203 10:13:40.989695 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:40 crc kubenswrapper[4756]: I0203 10:13:40.995743 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4gc6f"/"default-dockercfg-5d8m8" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.053426 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjgsp\" (UniqueName: \"kubernetes.io/projected/f2e63bb2-0743-4fcb-b686-114ddf15574e-kube-api-access-bjgsp\") pod \"crc-debug-mjkv4\" (UID: \"f2e63bb2-0743-4fcb-b686-114ddf15574e\") " pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.053854 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2e63bb2-0743-4fcb-b686-114ddf15574e-host\") pod \"crc-debug-mjkv4\" (UID: \"f2e63bb2-0743-4fcb-b686-114ddf15574e\") " pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.156144 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2e63bb2-0743-4fcb-b686-114ddf15574e-host\") pod \"crc-debug-mjkv4\" (UID: \"f2e63bb2-0743-4fcb-b686-114ddf15574e\") " pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.156280 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjgsp\" (UniqueName: \"kubernetes.io/projected/f2e63bb2-0743-4fcb-b686-114ddf15574e-kube-api-access-bjgsp\") pod \"crc-debug-mjkv4\" (UID: \"f2e63bb2-0743-4fcb-b686-114ddf15574e\") " pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.156659 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2e63bb2-0743-4fcb-b686-114ddf15574e-host\") pod \"crc-debug-mjkv4\" (UID: \"f2e63bb2-0743-4fcb-b686-114ddf15574e\") " pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.180554 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjgsp\" (UniqueName: \"kubernetes.io/projected/f2e63bb2-0743-4fcb-b686-114ddf15574e-kube-api-access-bjgsp\") pod \"crc-debug-mjkv4\" (UID: \"f2e63bb2-0743-4fcb-b686-114ddf15574e\") " pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.306997 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.627339 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da0736cb-84e1-46ad-853a-b4ac5b25afaa" path="/var/lib/kubelet/pods/da0736cb-84e1-46ad-853a-b4ac5b25afaa/volumes" Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.666597 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" event={"ID":"f2e63bb2-0743-4fcb-b686-114ddf15574e","Type":"ContainerStarted","Data":"f1f9e62bf3055aa2e82a1b674d32813a3265948a791c64bb23976f381eaa2dfa"} Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.666906 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" event={"ID":"f2e63bb2-0743-4fcb-b686-114ddf15574e","Type":"ContainerStarted","Data":"14ce03d156f2bd339e2c75ed623e66a95866b2ae4123f462156cb1911a624069"} Feb 03 10:13:41 crc kubenswrapper[4756]: I0203 10:13:41.680336 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" podStartSLOduration=1.6803207489999998 podStartE2EDuration="1.680320749s" podCreationTimestamp="2026-02-03 10:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-03 10:13:41.678295685 +0000 UTC m=+3812.828763060" watchObservedRunningTime="2026-02-03 10:13:41.680320749 +0000 UTC m=+3812.830788124" Feb 03 10:13:42 crc kubenswrapper[4756]: I0203 10:13:42.677161 4756 generic.go:334] "Generic (PLEG): container finished" podID="f2e63bb2-0743-4fcb-b686-114ddf15574e" containerID="f1f9e62bf3055aa2e82a1b674d32813a3265948a791c64bb23976f381eaa2dfa" exitCode=0 Feb 03 10:13:42 crc kubenswrapper[4756]: I0203 10:13:42.677216 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" event={"ID":"f2e63bb2-0743-4fcb-b686-114ddf15574e","Type":"ContainerDied","Data":"f1f9e62bf3055aa2e82a1b674d32813a3265948a791c64bb23976f381eaa2dfa"} Feb 03 10:13:43 crc kubenswrapper[4756]: I0203 10:13:43.784522 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:43 crc kubenswrapper[4756]: I0203 10:13:43.816952 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-mjkv4"] Feb 03 10:13:43 crc kubenswrapper[4756]: I0203 10:13:43.826155 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-mjkv4"] Feb 03 10:13:43 crc kubenswrapper[4756]: I0203 10:13:43.910876 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjgsp\" (UniqueName: \"kubernetes.io/projected/f2e63bb2-0743-4fcb-b686-114ddf15574e-kube-api-access-bjgsp\") pod \"f2e63bb2-0743-4fcb-b686-114ddf15574e\" (UID: \"f2e63bb2-0743-4fcb-b686-114ddf15574e\") " Feb 03 10:13:43 crc kubenswrapper[4756]: I0203 10:13:43.911182 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2e63bb2-0743-4fcb-b686-114ddf15574e-host\") pod \"f2e63bb2-0743-4fcb-b686-114ddf15574e\" (UID: \"f2e63bb2-0743-4fcb-b686-114ddf15574e\") " Feb 03 10:13:43 crc kubenswrapper[4756]: I0203 10:13:43.911378 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2e63bb2-0743-4fcb-b686-114ddf15574e-host" (OuterVolumeSpecName: "host") pod "f2e63bb2-0743-4fcb-b686-114ddf15574e" (UID: "f2e63bb2-0743-4fcb-b686-114ddf15574e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 10:13:43 crc kubenswrapper[4756]: I0203 10:13:43.911749 4756 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f2e63bb2-0743-4fcb-b686-114ddf15574e-host\") on node \"crc\" DevicePath \"\"" Feb 03 10:13:43 crc kubenswrapper[4756]: I0203 10:13:43.916542 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2e63bb2-0743-4fcb-b686-114ddf15574e-kube-api-access-bjgsp" (OuterVolumeSpecName: "kube-api-access-bjgsp") pod "f2e63bb2-0743-4fcb-b686-114ddf15574e" (UID: "f2e63bb2-0743-4fcb-b686-114ddf15574e"). InnerVolumeSpecName "kube-api-access-bjgsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:13:44 crc kubenswrapper[4756]: I0203 10:13:44.014120 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjgsp\" (UniqueName: \"kubernetes.io/projected/f2e63bb2-0743-4fcb-b686-114ddf15574e-kube-api-access-bjgsp\") on node \"crc\" DevicePath \"\"" Feb 03 10:13:44 crc kubenswrapper[4756]: I0203 10:13:44.694472 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14ce03d156f2bd339e2c75ed623e66a95866b2ae4123f462156cb1911a624069" Feb 03 10:13:44 crc kubenswrapper[4756]: I0203 10:13:44.694551 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-mjkv4" Feb 03 10:13:44 crc kubenswrapper[4756]: I0203 10:13:44.964117 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-xcxsx"] Feb 03 10:13:44 crc kubenswrapper[4756]: E0203 10:13:44.964599 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e63bb2-0743-4fcb-b686-114ddf15574e" containerName="container-00" Feb 03 10:13:44 crc kubenswrapper[4756]: I0203 10:13:44.964615 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e63bb2-0743-4fcb-b686-114ddf15574e" containerName="container-00" Feb 03 10:13:44 crc kubenswrapper[4756]: I0203 10:13:44.964857 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2e63bb2-0743-4fcb-b686-114ddf15574e" containerName="container-00" Feb 03 10:13:44 crc kubenswrapper[4756]: I0203 10:13:44.965805 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:44 crc kubenswrapper[4756]: I0203 10:13:44.968202 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4gc6f"/"default-dockercfg-5d8m8" Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.034085 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c61a4125-36f1-44a2-a3d9-1d0602d5c567-host\") pod \"crc-debug-xcxsx\" (UID: \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\") " pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.034867 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw7mp\" (UniqueName: \"kubernetes.io/projected/c61a4125-36f1-44a2-a3d9-1d0602d5c567-kube-api-access-tw7mp\") pod \"crc-debug-xcxsx\" (UID: \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\") " pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.136912 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c61a4125-36f1-44a2-a3d9-1d0602d5c567-host\") pod \"crc-debug-xcxsx\" (UID: \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\") " pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.136958 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw7mp\" (UniqueName: \"kubernetes.io/projected/c61a4125-36f1-44a2-a3d9-1d0602d5c567-kube-api-access-tw7mp\") pod \"crc-debug-xcxsx\" (UID: \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\") " pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.137065 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c61a4125-36f1-44a2-a3d9-1d0602d5c567-host\") pod \"crc-debug-xcxsx\" (UID: \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\") " pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.155961 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw7mp\" (UniqueName: \"kubernetes.io/projected/c61a4125-36f1-44a2-a3d9-1d0602d5c567-kube-api-access-tw7mp\") pod \"crc-debug-xcxsx\" (UID: \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\") " pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.281077 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:45 crc kubenswrapper[4756]: W0203 10:13:45.316812 4756 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc61a4125_36f1_44a2_a3d9_1d0602d5c567.slice/crio-1e36bf34849f3707d9eb5c1230b735bf50c3f53bde3fe06c661799a089a614cf WatchSource:0}: Error finding container 1e36bf34849f3707d9eb5c1230b735bf50c3f53bde3fe06c661799a089a614cf: Status 404 returned error can't find the container with id 1e36bf34849f3707d9eb5c1230b735bf50c3f53bde3fe06c661799a089a614cf Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.624931 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2e63bb2-0743-4fcb-b686-114ddf15574e" path="/var/lib/kubelet/pods/f2e63bb2-0743-4fcb-b686-114ddf15574e/volumes" Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.704390 4756 generic.go:334] "Generic (PLEG): container finished" podID="c61a4125-36f1-44a2-a3d9-1d0602d5c567" containerID="53514178268ee21df44ca9823c3533c42f0e78e4dddcd67d1caecf11c62fbf18" exitCode=0 Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.704532 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" event={"ID":"c61a4125-36f1-44a2-a3d9-1d0602d5c567","Type":"ContainerDied","Data":"53514178268ee21df44ca9823c3533c42f0e78e4dddcd67d1caecf11c62fbf18"} Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.704566 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" event={"ID":"c61a4125-36f1-44a2-a3d9-1d0602d5c567","Type":"ContainerStarted","Data":"1e36bf34849f3707d9eb5c1230b735bf50c3f53bde3fe06c661799a089a614cf"} Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.744936 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-xcxsx"] Feb 03 10:13:45 crc kubenswrapper[4756]: I0203 10:13:45.753068 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gc6f/crc-debug-xcxsx"] Feb 03 10:13:46 crc kubenswrapper[4756]: I0203 10:13:46.809857 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:13:46 crc kubenswrapper[4756]: I0203 10:13:46.974267 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw7mp\" (UniqueName: \"kubernetes.io/projected/c61a4125-36f1-44a2-a3d9-1d0602d5c567-kube-api-access-tw7mp\") pod \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\" (UID: \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\") " Feb 03 10:13:46 crc kubenswrapper[4756]: I0203 10:13:46.974500 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c61a4125-36f1-44a2-a3d9-1d0602d5c567-host\") pod \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\" (UID: \"c61a4125-36f1-44a2-a3d9-1d0602d5c567\") " Feb 03 10:13:46 crc kubenswrapper[4756]: I0203 10:13:46.974638 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c61a4125-36f1-44a2-a3d9-1d0602d5c567-host" (OuterVolumeSpecName: "host") pod "c61a4125-36f1-44a2-a3d9-1d0602d5c567" (UID: "c61a4125-36f1-44a2-a3d9-1d0602d5c567"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 03 10:13:46 crc kubenswrapper[4756]: I0203 10:13:46.976127 4756 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c61a4125-36f1-44a2-a3d9-1d0602d5c567-host\") on node \"crc\" DevicePath \"\"" Feb 03 10:13:46 crc kubenswrapper[4756]: I0203 10:13:46.985809 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61a4125-36f1-44a2-a3d9-1d0602d5c567-kube-api-access-tw7mp" (OuterVolumeSpecName: "kube-api-access-tw7mp") pod "c61a4125-36f1-44a2-a3d9-1d0602d5c567" (UID: "c61a4125-36f1-44a2-a3d9-1d0602d5c567"). InnerVolumeSpecName "kube-api-access-tw7mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:13:47 crc kubenswrapper[4756]: I0203 10:13:47.080167 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw7mp\" (UniqueName: \"kubernetes.io/projected/c61a4125-36f1-44a2-a3d9-1d0602d5c567-kube-api-access-tw7mp\") on node \"crc\" DevicePath \"\"" Feb 03 10:13:47 crc kubenswrapper[4756]: I0203 10:13:47.614539 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:13:47 crc kubenswrapper[4756]: E0203 10:13:47.614904 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:13:47 crc kubenswrapper[4756]: I0203 10:13:47.629718 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c61a4125-36f1-44a2-a3d9-1d0602d5c567" path="/var/lib/kubelet/pods/c61a4125-36f1-44a2-a3d9-1d0602d5c567/volumes" Feb 03 10:13:47 crc kubenswrapper[4756]: I0203 10:13:47.723213 4756 scope.go:117] "RemoveContainer" containerID="53514178268ee21df44ca9823c3533c42f0e78e4dddcd67d1caecf11c62fbf18" Feb 03 10:13:47 crc kubenswrapper[4756]: I0203 10:13:47.723318 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/crc-debug-xcxsx" Feb 03 10:14:02 crc kubenswrapper[4756]: I0203 10:14:02.613938 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:14:02 crc kubenswrapper[4756]: E0203 10:14:02.614640 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.103156 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f69df96f4-7db64_ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f/barbican-api/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.294465 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f69df96f4-7db64_ce3520f0-c9cd-41cc-ae0f-d0f7d9fa457f/barbican-api-log/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.347393 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5948f79bd-cl5cq_c34e66d1-0527-4f40-a676-1ac69c4d723f/barbican-keystone-listener/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.383493 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5948f79bd-cl5cq_c34e66d1-0527-4f40-a676-1ac69c4d723f/barbican-keystone-listener-log/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.516484 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68b68864ff-4znh2_8b049610-c8e6-40bd-98cb-138f42f1c35f/barbican-worker-log/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.530605 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68b68864ff-4znh2_8b049610-c8e6-40bd-98cb-138f42f1c35f/barbican-worker/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.813893 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_198484fd-46b3-4299-8938-f49fdfdcacb0/ceilometer-central-agent/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.828301 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fdlks_5e81e9f8-2003-4c8b-8270-bf0214a074e7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.892883 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_198484fd-46b3-4299-8938-f49fdfdcacb0/ceilometer-notification-agent/0.log" Feb 03 10:14:09 crc kubenswrapper[4756]: I0203 10:14:09.999889 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_198484fd-46b3-4299-8938-f49fdfdcacb0/proxy-httpd/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.044825 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_198484fd-46b3-4299-8938-f49fdfdcacb0/sg-core/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.135359 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26/cinder-api/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.200143 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1d26ce45-1b9c-4aad-ad26-6e43f5b8fe26/cinder-api-log/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.336600 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5abc83aa-9e60-4884-9633-d151104a151a/cinder-scheduler/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.376080 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5abc83aa-9e60-4884-9633-d151104a151a/probe/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.515301 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-dhvw2_3c14e04f-9c29-45ae-84dc-bad543ff55d2/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.563292 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-xtvjj_9851b3a8-4288-4c0b-9ae9-38a5c043682f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.721798 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-wtq86_2fa5e86f-f1ee-4e6e-9dba-26c7055a767e/init/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.892703 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-wtq86_2fa5e86f-f1ee-4e6e-9dba-26c7055a767e/init/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.917607 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-wtq86_2fa5e86f-f1ee-4e6e-9dba-26c7055a767e/dnsmasq-dns/0.log" Feb 03 10:14:10 crc kubenswrapper[4756]: I0203 10:14:10.941671 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-blt9q_2847ae70-333c-420d-b7a6-2c547477f35b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:11 crc kubenswrapper[4756]: I0203 10:14:11.115755 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2a921c4c-56ce-4689-b25d-0b4b69a01cdb/glance-log/0.log" Feb 03 10:14:11 crc kubenswrapper[4756]: I0203 10:14:11.136091 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2a921c4c-56ce-4689-b25d-0b4b69a01cdb/glance-httpd/0.log" Feb 03 10:14:11 crc kubenswrapper[4756]: I0203 10:14:11.289243 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bd8465a-e888-4b54-a570-76bcdc4736ac/glance-httpd/0.log" Feb 03 10:14:11 crc kubenswrapper[4756]: I0203 10:14:11.319639 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bd8465a-e888-4b54-a570-76bcdc4736ac/glance-log/0.log" Feb 03 10:14:11 crc kubenswrapper[4756]: I0203 10:14:11.462431 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d95f7b7bb-r6fd8_c34fbdee-8998-4e2d-bc98-c5c73ff35160/horizon/0.log" Feb 03 10:14:11 crc kubenswrapper[4756]: I0203 10:14:11.577183 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-9nclw_2396060c-bdc9-4f9c-8a42-c86984922418/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:11 crc kubenswrapper[4756]: I0203 10:14:11.821257 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-q5fk9_bf98eb35-81c9-4cc5-9d5c-d81d97d716ad/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:11 crc kubenswrapper[4756]: I0203 10:14:11.832552 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d95f7b7bb-r6fd8_c34fbdee-8998-4e2d-bc98-c5c73ff35160/horizon-log/0.log" Feb 03 10:14:12 crc kubenswrapper[4756]: I0203 10:14:12.040394 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-765cf46959-4rcst_37cb5fd2-29ae-4c37-9ca5-f2c5aa3240bf/keystone-api/0.log" Feb 03 10:14:12 crc kubenswrapper[4756]: I0203 10:14:12.079968 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29501881-dszb5_bccf9d07-df25-4256-8a89-f2a47ee20a73/keystone-cron/0.log" Feb 03 10:14:12 crc kubenswrapper[4756]: I0203 10:14:12.232542 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_dc0c1050-2088-4a2d-98bc-fb14078d6418/kube-state-metrics/0.log" Feb 03 10:14:12 crc kubenswrapper[4756]: I0203 10:14:12.313973 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-kzq8m_9419b87e-a027-4476-a95b-e4bc884d90e7/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:12 crc kubenswrapper[4756]: I0203 10:14:12.644240 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8547fc4b5c-w4sjb_3a5ca842-66f7-4ff0-b8f7-afe33276b34b/neutron-api/0.log" Feb 03 10:14:12 crc kubenswrapper[4756]: I0203 10:14:12.784510 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8547fc4b5c-w4sjb_3a5ca842-66f7-4ff0-b8f7-afe33276b34b/neutron-httpd/0.log" Feb 03 10:14:12 crc kubenswrapper[4756]: I0203 10:14:12.889699 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-j8c2f_cebe42d5-5283-4c1e-a2f2-70db6d3e2257/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:13 crc kubenswrapper[4756]: I0203 10:14:13.463549 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ca27353e-8cea-4f67-a90b-9776e510e82e/nova-api-log/0.log" Feb 03 10:14:13 crc kubenswrapper[4756]: I0203 10:14:13.562315 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_84eaadd5-45a9-4055-a567-9e72fe3141b2/nova-cell0-conductor-conductor/0.log" Feb 03 10:14:13 crc kubenswrapper[4756]: I0203 10:14:13.747519 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ca27353e-8cea-4f67-a90b-9776e510e82e/nova-api-api/0.log" Feb 03 10:14:13 crc kubenswrapper[4756]: I0203 10:14:13.828723 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_7978abb6-3f13-4b6e-8300-38a9d9cdd335/nova-cell1-conductor-conductor/0.log" Feb 03 10:14:13 crc kubenswrapper[4756]: I0203 10:14:13.895806 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_58183079-d9ac-4407-bec4-e93605aa884d/nova-cell1-novncproxy-novncproxy/0.log" Feb 03 10:14:14 crc kubenswrapper[4756]: I0203 10:14:14.028558 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-k5dsp_0573a518-65e1-4717-a24e-285e30cad247/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:14 crc kubenswrapper[4756]: I0203 10:14:14.231512 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_776f1390-6de5-4054-bff2-58621c198147/nova-metadata-log/0.log" Feb 03 10:14:14 crc kubenswrapper[4756]: I0203 10:14:14.511796 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c/mysql-bootstrap/0.log" Feb 03 10:14:14 crc kubenswrapper[4756]: I0203 10:14:14.524321 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_fead6bc0-9b80-4ab0-938c-19fad62f8178/nova-scheduler-scheduler/0.log" Feb 03 10:14:14 crc kubenswrapper[4756]: I0203 10:14:14.682348 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c/mysql-bootstrap/0.log" Feb 03 10:14:14 crc kubenswrapper[4756]: I0203 10:14:14.755363 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_47af8efb-b1fa-4a7b-be26-6b35f2ba9c2c/galera/0.log" Feb 03 10:14:14 crc kubenswrapper[4756]: I0203 10:14:14.931202 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_77baa89d-5260-4b1d-a633-a4f3e37ed6bb/mysql-bootstrap/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.105115 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_77baa89d-5260-4b1d-a633-a4f3e37ed6bb/mysql-bootstrap/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.139521 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_77baa89d-5260-4b1d-a633-a4f3e37ed6bb/galera/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.305927 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3e4a2949-6bbf-4478-aee7-2cd98095c059/openstackclient/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.386991 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-d4749_f6c5f3c6-ec4f-4c2a-beb0-128e53ecb89a/openstack-network-exporter/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.537827 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_776f1390-6de5-4054-bff2-58621c198147/nova-metadata-metadata/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.585014 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb566_4600dd3a-bdd0-44b1-8a6d-f34313d5df5d/ovsdb-server-init/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.717052 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb566_4600dd3a-bdd0-44b1-8a6d-f34313d5df5d/ovsdb-server-init/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.747900 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb566_4600dd3a-bdd0-44b1-8a6d-f34313d5df5d/ovs-vswitchd/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.818260 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fb566_4600dd3a-bdd0-44b1-8a6d-f34313d5df5d/ovsdb-server/0.log" Feb 03 10:14:15 crc kubenswrapper[4756]: I0203 10:14:15.894709 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-wnf95_15695b6e-11fa-4913-b502-a78974ddf81c/ovn-controller/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.053047 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-jbd5l_3c6bade3-63d5-4342-8ded-feaf11b54263/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.156223 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_912cca3d-0179-4199-b236-a6ae37c9a80d/openstack-network-exporter/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.214192 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_912cca3d-0179-4199-b236-a6ae37c9a80d/ovn-northd/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.361565 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a2da6bae-fd2d-4f9c-b2aa-440d4264f17a/openstack-network-exporter/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.415261 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a2da6bae-fd2d-4f9c-b2aa-440d4264f17a/ovsdbserver-nb/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.587958 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bfba7eb1-da32-4bc6-8309-f076ad7b32a0/ovsdbserver-sb/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.591064 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bfba7eb1-da32-4bc6-8309-f076ad7b32a0/openstack-network-exporter/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.613805 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:14:16 crc kubenswrapper[4756]: E0203 10:14:16.614179 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.856326 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f5fdf688d-jsswq_5446426b-adaf-4af7-97f8-b8b2d798d364/placement-api/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.908908 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60b10ab9-d28a-4c83-9e67-02d1499685ce/setup-container/0.log" Feb 03 10:14:16 crc kubenswrapper[4756]: I0203 10:14:16.929171 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f5fdf688d-jsswq_5446426b-adaf-4af7-97f8-b8b2d798d364/placement-log/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.113946 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60b10ab9-d28a-4c83-9e67-02d1499685ce/setup-container/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.152753 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_60b10ab9-d28a-4c83-9e67-02d1499685ce/rabbitmq/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.195090 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f76f412-ff7c-4857-9670-4721dc5c3080/setup-container/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.428041 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-frvhw_4ae5e9e1-d148-4487-862a-b2312222f2b0/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.434944 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f76f412-ff7c-4857-9670-4721dc5c3080/setup-container/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.478291 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f76f412-ff7c-4857-9670-4721dc5c3080/rabbitmq/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.676742 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-xvkxf_ce26979b-e0a6-41a5-b56d-e1800d5d6704/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.745682 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7c64v_200f01ff-d33a-474b-944e-e6ed2486db68/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:17 crc kubenswrapper[4756]: I0203 10:14:17.923467 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9g7tk_5181cb59-0c71-4a78-8ba8-d5c5a22c4dbf/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:18 crc kubenswrapper[4756]: I0203 10:14:18.013130 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-7jsx4_848ce5a7-c89c-4cac-8275-263ca14eac78/ssh-known-hosts-edpm-deployment/0.log" Feb 03 10:14:18 crc kubenswrapper[4756]: I0203 10:14:18.342489 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85ff9cc7c5-ns7sz_74be27bd-98ba-4f2c-bd2e-ea5855812a55/proxy-httpd/0.log" Feb 03 10:14:18 crc kubenswrapper[4756]: I0203 10:14:18.632783 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-85ff9cc7c5-ns7sz_74be27bd-98ba-4f2c-bd2e-ea5855812a55/proxy-server/0.log" Feb 03 10:14:18 crc kubenswrapper[4756]: I0203 10:14:18.689771 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-pqfmb_f356b5c0-fae0-4a67-9624-4fbe59bd2f9e/swift-ring-rebalance/0.log" Feb 03 10:14:18 crc kubenswrapper[4756]: I0203 10:14:18.808388 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/account-auditor/0.log" Feb 03 10:14:18 crc kubenswrapper[4756]: I0203 10:14:18.873211 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/account-reaper/0.log" Feb 03 10:14:18 crc kubenswrapper[4756]: I0203 10:14:18.976704 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/account-replicator/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.053221 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/account-server/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.120964 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/container-replicator/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.128583 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/container-auditor/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.188975 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/container-server/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.248527 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/container-updater/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.350191 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-auditor/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.375620 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-expirer/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.400376 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-replicator/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.481771 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-server/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.551203 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/rsync/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.577590 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/object-updater/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.584493 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d4dd6c35-7256-4f4b-9ef7-f09b8ff9dda9/swift-recon-cron/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.859362 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_c529c965-e4c3-4e32-a5b2-445aebef0f08/tempest-tests-tempest-tests-runner/0.log" Feb 03 10:14:19 crc kubenswrapper[4756]: I0203 10:14:19.864727 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-v4bzd_17726fdd-5344-437e-8060-5e26bcfb28cd/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:20 crc kubenswrapper[4756]: I0203 10:14:20.089212 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_e5cc0402-4e5a-49ab-bd96-654e19fe4048/test-operator-logs-container/0.log" Feb 03 10:14:20 crc kubenswrapper[4756]: I0203 10:14:20.203264 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wdbcc_4a58ee0e-41e0-4040-99c1-3261c3416a29/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 03 10:14:29 crc kubenswrapper[4756]: I0203 10:14:29.629126 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:14:29 crc kubenswrapper[4756]: E0203 10:14:29.630030 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:14:30 crc kubenswrapper[4756]: I0203 10:14:30.007960 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ce92a835-f667-4f5b-b9e1-c8afcf476b9a/memcached/0.log" Feb 03 10:14:42 crc kubenswrapper[4756]: I0203 10:14:42.614414 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:14:42 crc kubenswrapper[4756]: E0203 10:14:42.615053 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.099438 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/util/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.250916 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/util/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.283304 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/pull/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.303142 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/pull/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.452567 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/util/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.477087 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/pull/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.482331 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_198fefeaf9ba73d70a516f9552fef45f45d71dbe2e17427e2d24935a28chpt8_1e21c47a-ab4b-4663-ba66-21f3b8c4975b/extract/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.714478 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d874c8fc-dplk2_cbbe13dd-d5aa-4e9f-b901-2a135c8cfdce/manager/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.747807 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b6c4d8c5f-6pbkt_a6153ce6-006d-4351-8a9b-aa0abe1cbdad/manager/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.862393 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d9697b7f4-qnklz_d74f0516-3682-4b80-beb4-f7237c4daff8/manager/0.log" Feb 03 10:14:44 crc kubenswrapper[4756]: I0203 10:14:44.979052 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8886f4c47-pgdzr_6dd1801b-ac57-4d28-9554-ef8ad7148c08/manager/0.log" Feb 03 10:14:45 crc kubenswrapper[4756]: I0203 10:14:45.050785 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69d6db494d-564mj_63c547ee-bb8c-4f1a-b29a-12a80c1749f4/manager/0.log" Feb 03 10:14:45 crc kubenswrapper[4756]: I0203 10:14:45.195085 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5fb775575f-hvnx5_d9f68e89-0c29-4d3d-9dae-7448ab35206f/manager/0.log" Feb 03 10:14:45 crc kubenswrapper[4756]: I0203 10:14:45.446935 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f4b8bd54d-zhlxl_ad043c5b-465c-43b2-b73c-45e605fc7b25/manager/0.log" Feb 03 10:14:45 crc kubenswrapper[4756]: I0203 10:14:45.562426 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79955696d6-vvkmq_aae19bbd-e8ec-4a6f-8e77-9dd46373e31b/manager/0.log" Feb 03 10:14:45 crc kubenswrapper[4756]: I0203 10:14:45.655721 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-84f48565d4-w7vdf_919fda51-9a13-4233-ba62-a85ac3559a02/manager/0.log" Feb 03 10:14:45 crc kubenswrapper[4756]: I0203 10:14:45.705174 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7dd968899f-jwlpw_d8ad4baf-2108-4286-a2f6-5a8ab537185c/manager/0.log" Feb 03 10:14:45 crc kubenswrapper[4756]: I0203 10:14:45.846755 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf948998-5q5s7_aacb5f03-cd60-4589-9758-9fb4e1b68030/manager/0.log" Feb 03 10:14:45 crc kubenswrapper[4756]: I0203 10:14:45.929840 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-585dbc889-gn2d2_4483922a-5bb6-40da-bb78-cdeca488ccd8/manager/0.log" Feb 03 10:14:46 crc kubenswrapper[4756]: I0203 10:14:46.110076 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-55bff696bd-qnwj2_999c5dd9-6dbb-44b9-bdd7-52c146d34e87/manager/0.log" Feb 03 10:14:46 crc kubenswrapper[4756]: I0203 10:14:46.148263 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6687f8d877-bqthj_932503a4-3932-4127-a2b2-a2e0291782b7/manager/0.log" Feb 03 10:14:46 crc kubenswrapper[4756]: I0203 10:14:46.320357 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-59c4b45c4dssdbx_a0ebd520-4c15-4c12-89fe-2eb8a407ca25/manager/0.log" Feb 03 10:14:46 crc kubenswrapper[4756]: I0203 10:14:46.419410 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-7f67889c79-dwpgg_2e4a4670-318d-4489-9a71-a084e2445361/operator/0.log" Feb 03 10:14:46 crc kubenswrapper[4756]: I0203 10:14:46.599279 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-r84l6_9dc52ada-64d7-45a9-b880-ded26c6ec2f6/registry-server/0.log" Feb 03 10:14:46 crc kubenswrapper[4756]: I0203 10:14:46.836559 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-788c46999f-wl9ft_bd1225c5-fbb5-40f5-8e71-6c1075feeb16/manager/0.log" Feb 03 10:14:46 crc kubenswrapper[4756]: I0203 10:14:46.928699 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b964cf4cd-7rrf6_58c01e66-fe0e-45f2-8caf-df41e1dad607/manager/0.log" Feb 03 10:14:47 crc kubenswrapper[4756]: I0203 10:14:47.186502 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-lqghx_9fe0931c-fae1-4dfd-968d-bccf61682fa5/operator/0.log" Feb 03 10:14:47 crc kubenswrapper[4756]: I0203 10:14:47.324710 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68fc8c869-w87dt_51a902ff-23e6-4578-809f-0de9a6c5cbae/manager/0.log" Feb 03 10:14:47 crc kubenswrapper[4756]: I0203 10:14:47.553996 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64b5b76f97-2cw2v_d7f70a3a-d4db-4d12-8838-9d8923f06213/manager/0.log" Feb 03 10:14:47 crc kubenswrapper[4756]: I0203 10:14:47.634513 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c6ffddb96-rvk7w_2183ff26-c37d-472e-bf53-e0beba79a31d/manager/0.log" Feb 03 10:14:47 crc kubenswrapper[4756]: I0203 10:14:47.652333 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56f8bfcd9f-l5wms_04edde9e-f443-4e9e-929b-65e2d9258ff2/manager/0.log" Feb 03 10:14:47 crc kubenswrapper[4756]: I0203 10:14:47.788813 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-9rdzw_ef6310fb-d44c-449f-b3d0-72c20e4b3d7b/manager/0.log" Feb 03 10:14:55 crc kubenswrapper[4756]: I0203 10:14:55.613980 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:14:56 crc kubenswrapper[4756]: I0203 10:14:56.311698 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"659a44e09cf2712199a26e5ab73d9a728d8d93116f1d5994260fd77abcbbfc49"} Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.179623 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw"] Feb 03 10:15:00 crc kubenswrapper[4756]: E0203 10:15:00.182476 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61a4125-36f1-44a2-a3d9-1d0602d5c567" containerName="container-00" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.182507 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61a4125-36f1-44a2-a3d9-1d0602d5c567" containerName="container-00" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.182742 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61a4125-36f1-44a2-a3d9-1d0602d5c567" containerName="container-00" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.183418 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.186549 4756 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.186635 4756 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.203799 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw"] Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.306630 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d70ec-9e00-4db3-82fa-40b693ca9379-config-volume\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.306701 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6bfr\" (UniqueName: \"kubernetes.io/projected/f68d70ec-9e00-4db3-82fa-40b693ca9379-kube-api-access-k6bfr\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.306914 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d70ec-9e00-4db3-82fa-40b693ca9379-secret-volume\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.409152 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d70ec-9e00-4db3-82fa-40b693ca9379-secret-volume\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.409332 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d70ec-9e00-4db3-82fa-40b693ca9379-config-volume\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.409374 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6bfr\" (UniqueName: \"kubernetes.io/projected/f68d70ec-9e00-4db3-82fa-40b693ca9379-kube-api-access-k6bfr\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.410346 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d70ec-9e00-4db3-82fa-40b693ca9379-config-volume\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.416685 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d70ec-9e00-4db3-82fa-40b693ca9379-secret-volume\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.428127 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6bfr\" (UniqueName: \"kubernetes.io/projected/f68d70ec-9e00-4db3-82fa-40b693ca9379-kube-api-access-k6bfr\") pod \"collect-profiles-29501895-6k9fw\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.509236 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:00 crc kubenswrapper[4756]: I0203 10:15:00.961614 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw"] Feb 03 10:15:01 crc kubenswrapper[4756]: I0203 10:15:01.351365 4756 generic.go:334] "Generic (PLEG): container finished" podID="f68d70ec-9e00-4db3-82fa-40b693ca9379" containerID="39e2fd3cec40300a3a274a4435ff64f8b0fcf1d0ef31eec82a1f4eef2cd57220" exitCode=0 Feb 03 10:15:01 crc kubenswrapper[4756]: I0203 10:15:01.351631 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" event={"ID":"f68d70ec-9e00-4db3-82fa-40b693ca9379","Type":"ContainerDied","Data":"39e2fd3cec40300a3a274a4435ff64f8b0fcf1d0ef31eec82a1f4eef2cd57220"} Feb 03 10:15:01 crc kubenswrapper[4756]: I0203 10:15:01.351738 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" event={"ID":"f68d70ec-9e00-4db3-82fa-40b693ca9379","Type":"ContainerStarted","Data":"6726b5bbea6effd1bb1e6de3f8acced1616a7e2e8509adf185f5e1c278049484"} Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.757377 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.863784 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6bfr\" (UniqueName: \"kubernetes.io/projected/f68d70ec-9e00-4db3-82fa-40b693ca9379-kube-api-access-k6bfr\") pod \"f68d70ec-9e00-4db3-82fa-40b693ca9379\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.863889 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d70ec-9e00-4db3-82fa-40b693ca9379-secret-volume\") pod \"f68d70ec-9e00-4db3-82fa-40b693ca9379\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.864008 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d70ec-9e00-4db3-82fa-40b693ca9379-config-volume\") pod \"f68d70ec-9e00-4db3-82fa-40b693ca9379\" (UID: \"f68d70ec-9e00-4db3-82fa-40b693ca9379\") " Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.864923 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68d70ec-9e00-4db3-82fa-40b693ca9379-config-volume" (OuterVolumeSpecName: "config-volume") pod "f68d70ec-9e00-4db3-82fa-40b693ca9379" (UID: "f68d70ec-9e00-4db3-82fa-40b693ca9379"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.870054 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f68d70ec-9e00-4db3-82fa-40b693ca9379-kube-api-access-k6bfr" (OuterVolumeSpecName: "kube-api-access-k6bfr") pod "f68d70ec-9e00-4db3-82fa-40b693ca9379" (UID: "f68d70ec-9e00-4db3-82fa-40b693ca9379"). InnerVolumeSpecName "kube-api-access-k6bfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.870602 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f68d70ec-9e00-4db3-82fa-40b693ca9379-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f68d70ec-9e00-4db3-82fa-40b693ca9379" (UID: "f68d70ec-9e00-4db3-82fa-40b693ca9379"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.965810 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6bfr\" (UniqueName: \"kubernetes.io/projected/f68d70ec-9e00-4db3-82fa-40b693ca9379-kube-api-access-k6bfr\") on node \"crc\" DevicePath \"\"" Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.966151 4756 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f68d70ec-9e00-4db3-82fa-40b693ca9379-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 03 10:15:02 crc kubenswrapper[4756]: I0203 10:15:02.966223 4756 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f68d70ec-9e00-4db3-82fa-40b693ca9379-config-volume\") on node \"crc\" DevicePath \"\"" Feb 03 10:15:03 crc kubenswrapper[4756]: I0203 10:15:03.369930 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" event={"ID":"f68d70ec-9e00-4db3-82fa-40b693ca9379","Type":"ContainerDied","Data":"6726b5bbea6effd1bb1e6de3f8acced1616a7e2e8509adf185f5e1c278049484"} Feb 03 10:15:03 crc kubenswrapper[4756]: I0203 10:15:03.369977 4756 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6726b5bbea6effd1bb1e6de3f8acced1616a7e2e8509adf185f5e1c278049484" Feb 03 10:15:03 crc kubenswrapper[4756]: I0203 10:15:03.370370 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29501895-6k9fw" Feb 03 10:15:03 crc kubenswrapper[4756]: I0203 10:15:03.851882 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp"] Feb 03 10:15:03 crc kubenswrapper[4756]: I0203 10:15:03.862634 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29501850-5m9bp"] Feb 03 10:15:05 crc kubenswrapper[4756]: I0203 10:15:05.238528 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qzff8_641583b7-728e-4e81-ad79-efcc524c5a10/control-plane-machine-set-operator/0.log" Feb 03 10:15:05 crc kubenswrapper[4756]: I0203 10:15:05.406636 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bk8nb_f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e/kube-rbac-proxy/0.log" Feb 03 10:15:05 crc kubenswrapper[4756]: I0203 10:15:05.414081 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bk8nb_f98e53dc-7bca-4a0f-a614-2ffbf3efdf9e/machine-api-operator/0.log" Feb 03 10:15:05 crc kubenswrapper[4756]: I0203 10:15:05.647155 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abe4717b-f308-4623-8be9-bffc1097815a" path="/var/lib/kubelet/pods/abe4717b-f308-4623-8be9-bffc1097815a/volumes" Feb 03 10:15:18 crc kubenswrapper[4756]: I0203 10:15:18.697656 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-fm5vv_e8c43b40-0d26-4582-8ffd-b2387527edd6/cert-manager-controller/0.log" Feb 03 10:15:18 crc kubenswrapper[4756]: I0203 10:15:18.833479 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-gnvfj_2c537fcd-648d-4fea-ab9a-0e0edfa8b14b/cert-manager-cainjector/0.log" Feb 03 10:15:18 crc kubenswrapper[4756]: I0203 10:15:18.883692 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-gnr5k_a4b8e083-79fc-4bbe-b3c8-16e237b63458/cert-manager-webhook/0.log" Feb 03 10:15:31 crc kubenswrapper[4756]: I0203 10:15:31.455245 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-zscxn_ac470c2b-de36-4daa-bce0-8163e27e43e7/nmstate-console-plugin/0.log" Feb 03 10:15:31 crc kubenswrapper[4756]: I0203 10:15:31.572964 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fv282_e827e9ff-5a84-42cc-9927-56904973c3ad/nmstate-handler/0.log" Feb 03 10:15:31 crc kubenswrapper[4756]: I0203 10:15:31.705188 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-75x2d_63b70a58-8d39-418a-bedc-93f9a94540a4/kube-rbac-proxy/0.log" Feb 03 10:15:31 crc kubenswrapper[4756]: I0203 10:15:31.769359 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-75x2d_63b70a58-8d39-418a-bedc-93f9a94540a4/nmstate-metrics/0.log" Feb 03 10:15:31 crc kubenswrapper[4756]: I0203 10:15:31.884080 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-wvckj_6374dff8-08c2-49fe-b5f4-0eb5142f61bf/nmstate-operator/0.log" Feb 03 10:15:31 crc kubenswrapper[4756]: I0203 10:15:31.980476 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-cjp9b_b9dfc50f-d7ab-4806-aac1-70e36788c60e/nmstate-webhook/0.log" Feb 03 10:15:32 crc kubenswrapper[4756]: I0203 10:15:32.324039 4756 scope.go:117] "RemoveContainer" containerID="281354e82600aa1b0aecb578ffbf58ed7afd8326567ce5bab438bcf72c9dcdc2" Feb 03 10:16:00 crc kubenswrapper[4756]: I0203 10:16:00.409937 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-dc6ld_16bed8ad-994f-4826-b43d-bc8ef79bce7e/kube-rbac-proxy/0.log" Feb 03 10:16:00 crc kubenswrapper[4756]: I0203 10:16:00.463794 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-dc6ld_16bed8ad-994f-4826-b43d-bc8ef79bce7e/controller/0.log" Feb 03 10:16:00 crc kubenswrapper[4756]: I0203 10:16:00.586714 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-frr-files/0.log" Feb 03 10:16:00 crc kubenswrapper[4756]: I0203 10:16:00.837682 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-frr-files/0.log" Feb 03 10:16:00 crc kubenswrapper[4756]: I0203 10:16:00.855650 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-reloader/0.log" Feb 03 10:16:00 crc kubenswrapper[4756]: I0203 10:16:00.860035 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-metrics/0.log" Feb 03 10:16:00 crc kubenswrapper[4756]: I0203 10:16:00.906465 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-reloader/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.120995 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-reloader/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.147180 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-frr-files/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.175115 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-metrics/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.184880 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-metrics/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.324003 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-reloader/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.374248 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-metrics/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.385782 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/cp-frr-files/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.429847 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/controller/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.771681 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/frr-metrics/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.807991 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/kube-rbac-proxy/0.log" Feb 03 10:16:01 crc kubenswrapper[4756]: I0203 10:16:01.823548 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/kube-rbac-proxy-frr/0.log" Feb 03 10:16:02 crc kubenswrapper[4756]: I0203 10:16:02.032434 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/reloader/0.log" Feb 03 10:16:02 crc kubenswrapper[4756]: I0203 10:16:02.088658 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-bwmnx_bcaa1bc4-f6a5-4216-8581-13cbf07c2504/frr-k8s-webhook-server/0.log" Feb 03 10:16:02 crc kubenswrapper[4756]: I0203 10:16:02.340108 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-576684d48c-mkg4k_fc59b7e7-12f3-4501-94f9-07f3338d7835/manager/0.log" Feb 03 10:16:02 crc kubenswrapper[4756]: I0203 10:16:02.483988 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-87484dcc8-cbd6x_12aaddd9-3223-4da5-8040-4149116fede4/webhook-server/0.log" Feb 03 10:16:02 crc kubenswrapper[4756]: I0203 10:16:02.609957 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-44db6_98d5a2ec-16f8-4654-b5ba-0c32b08d0fda/kube-rbac-proxy/0.log" Feb 03 10:16:03 crc kubenswrapper[4756]: I0203 10:16:03.154298 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-88fxn_c4e946d8-45d6-4706-982c-32d811489a5a/frr/0.log" Feb 03 10:16:03 crc kubenswrapper[4756]: I0203 10:16:03.218748 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-44db6_98d5a2ec-16f8-4654-b5ba-0c32b08d0fda/speaker/0.log" Feb 03 10:16:17 crc kubenswrapper[4756]: I0203 10:16:17.613076 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/util/0.log" Feb 03 10:16:17 crc kubenswrapper[4756]: I0203 10:16:17.854217 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/pull/0.log" Feb 03 10:16:17 crc kubenswrapper[4756]: I0203 10:16:17.856570 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/pull/0.log" Feb 03 10:16:17 crc kubenswrapper[4756]: I0203 10:16:17.950254 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/util/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.110830 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/pull/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.157297 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/util/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.174949 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dck4l5q_3dd53027-4438-4022-9ee5-8ed46a8537e3/extract/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.399493 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/util/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.552992 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/util/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.564159 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/pull/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.600747 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/pull/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.790116 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/util/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.811046 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/extract/0.log" Feb 03 10:16:18 crc kubenswrapper[4756]: I0203 10:16:18.811654 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713c5tp7_5db72b48-6cac-4211-a692-0e5d98e6d9e8/pull/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.008526 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-utilities/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.189061 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-utilities/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.201508 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-content/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.227821 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-content/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.421625 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-content/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.474033 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/extract-utilities/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.696532 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-utilities/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.924687 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kzdvb_74a5cfb6-5bda-4b5b-a828-1bf7a2705273/registry-server/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.929550 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-utilities/0.log" Feb 03 10:16:19 crc kubenswrapper[4756]: I0203 10:16:19.929679 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-content/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.165522 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-content/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.311746 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-utilities/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.391403 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/extract-content/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.584108 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xztlg_29956f45-fbd4-4630-82fe-8253dc497ce6/marketplace-operator/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.627855 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-utilities/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.783404 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2nqkf_f507b140-0d1f-424b-8077-80c7ac9a1a15/registry-server/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.902043 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-content/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.903943 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-utilities/0.log" Feb 03 10:16:20 crc kubenswrapper[4756]: I0203 10:16:20.984757 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-content/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.094476 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-utilities/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.152314 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/extract-content/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.264319 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-h5999_3021f6bf-d879-49e1-8051-9f599be91475/registry-server/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.335291 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-utilities/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.574678 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-content/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.579121 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-utilities/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.581973 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-content/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.768726 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-utilities/0.log" Feb 03 10:16:21 crc kubenswrapper[4756]: I0203 10:16:21.823108 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/extract-content/0.log" Feb 03 10:16:22 crc kubenswrapper[4756]: I0203 10:16:22.368306 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dvqhq_4ad66a22-bd41-40e4-b88f-d222ff093a15/registry-server/0.log" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.291774 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-27x92"] Feb 03 10:16:41 crc kubenswrapper[4756]: E0203 10:16:41.292761 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f68d70ec-9e00-4db3-82fa-40b693ca9379" containerName="collect-profiles" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.292778 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="f68d70ec-9e00-4db3-82fa-40b693ca9379" containerName="collect-profiles" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.293014 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="f68d70ec-9e00-4db3-82fa-40b693ca9379" containerName="collect-profiles" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.296025 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.311311 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-27x92"] Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.384040 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-catalog-content\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.384317 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-utilities\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.384475 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb96s\" (UniqueName: \"kubernetes.io/projected/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-kube-api-access-cb96s\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.486418 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb96s\" (UniqueName: \"kubernetes.io/projected/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-kube-api-access-cb96s\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.486545 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-catalog-content\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.486572 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-utilities\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.487193 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-utilities\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.487193 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-catalog-content\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.515715 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb96s\" (UniqueName: \"kubernetes.io/projected/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-kube-api-access-cb96s\") pod \"redhat-operators-27x92\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:41 crc kubenswrapper[4756]: I0203 10:16:41.615903 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:42 crc kubenswrapper[4756]: I0203 10:16:42.137225 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-27x92"] Feb 03 10:16:42 crc kubenswrapper[4756]: I0203 10:16:42.261545 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27x92" event={"ID":"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42","Type":"ContainerStarted","Data":"4b13826e640379092bec4694caca4597d2cf711ee8d2942282160b728d22acf6"} Feb 03 10:16:43 crc kubenswrapper[4756]: I0203 10:16:43.269593 4756 generic.go:334] "Generic (PLEG): container finished" podID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerID="2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6" exitCode=0 Feb 03 10:16:43 crc kubenswrapper[4756]: I0203 10:16:43.269702 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27x92" event={"ID":"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42","Type":"ContainerDied","Data":"2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6"} Feb 03 10:16:43 crc kubenswrapper[4756]: I0203 10:16:43.271554 4756 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 03 10:16:44 crc kubenswrapper[4756]: I0203 10:16:44.280599 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27x92" event={"ID":"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42","Type":"ContainerStarted","Data":"46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88"} Feb 03 10:16:47 crc kubenswrapper[4756]: I0203 10:16:47.305380 4756 generic.go:334] "Generic (PLEG): container finished" podID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerID="46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88" exitCode=0 Feb 03 10:16:47 crc kubenswrapper[4756]: I0203 10:16:47.305476 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27x92" event={"ID":"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42","Type":"ContainerDied","Data":"46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88"} Feb 03 10:16:49 crc kubenswrapper[4756]: I0203 10:16:49.324215 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27x92" event={"ID":"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42","Type":"ContainerStarted","Data":"ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3"} Feb 03 10:16:49 crc kubenswrapper[4756]: I0203 10:16:49.355100 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-27x92" podStartSLOduration=2.883629098 podStartE2EDuration="8.355079419s" podCreationTimestamp="2026-02-03 10:16:41 +0000 UTC" firstStartedPulling="2026-02-03 10:16:43.27127313 +0000 UTC m=+3994.421740505" lastFinishedPulling="2026-02-03 10:16:48.742723441 +0000 UTC m=+3999.893190826" observedRunningTime="2026-02-03 10:16:49.351196268 +0000 UTC m=+4000.501663643" watchObservedRunningTime="2026-02-03 10:16:49.355079419 +0000 UTC m=+4000.505546794" Feb 03 10:16:51 crc kubenswrapper[4756]: I0203 10:16:51.629614 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:51 crc kubenswrapper[4756]: I0203 10:16:51.631025 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:16:52 crc kubenswrapper[4756]: I0203 10:16:52.665939 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-27x92" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="registry-server" probeResult="failure" output=< Feb 03 10:16:52 crc kubenswrapper[4756]: timeout: failed to connect service ":50051" within 1s Feb 03 10:16:52 crc kubenswrapper[4756]: > Feb 03 10:17:02 crc kubenswrapper[4756]: I0203 10:17:02.692400 4756 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-27x92" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="registry-server" probeResult="failure" output=< Feb 03 10:17:02 crc kubenswrapper[4756]: timeout: failed to connect service ":50051" within 1s Feb 03 10:17:02 crc kubenswrapper[4756]: > Feb 03 10:17:11 crc kubenswrapper[4756]: I0203 10:17:11.676375 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:17:11 crc kubenswrapper[4756]: I0203 10:17:11.723174 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:17:12 crc kubenswrapper[4756]: I0203 10:17:12.490417 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-27x92"] Feb 03 10:17:13 crc kubenswrapper[4756]: I0203 10:17:13.547747 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-27x92" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="registry-server" containerID="cri-o://ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3" gracePeriod=2 Feb 03 10:17:13 crc kubenswrapper[4756]: I0203 10:17:13.565883 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:17:13 crc kubenswrapper[4756]: I0203 10:17:13.565955 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.030598 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.146707 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-catalog-content\") pod \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.146850 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb96s\" (UniqueName: \"kubernetes.io/projected/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-kube-api-access-cb96s\") pod \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.146913 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-utilities\") pod \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\" (UID: \"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42\") " Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.147832 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-utilities" (OuterVolumeSpecName: "utilities") pod "ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" (UID: "ef90a0a3-2ed0-494f-9278-2bccf7ea1c42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.153030 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-kube-api-access-cb96s" (OuterVolumeSpecName: "kube-api-access-cb96s") pod "ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" (UID: "ef90a0a3-2ed0-494f-9278-2bccf7ea1c42"). InnerVolumeSpecName "kube-api-access-cb96s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.249130 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb96s\" (UniqueName: \"kubernetes.io/projected/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-kube-api-access-cb96s\") on node \"crc\" DevicePath \"\"" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.249172 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.271236 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" (UID: "ef90a0a3-2ed0-494f-9278-2bccf7ea1c42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.351389 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.556999 4756 generic.go:334] "Generic (PLEG): container finished" podID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerID="ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3" exitCode=0 Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.557046 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27x92" event={"ID":"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42","Type":"ContainerDied","Data":"ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3"} Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.557073 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27x92" event={"ID":"ef90a0a3-2ed0-494f-9278-2bccf7ea1c42","Type":"ContainerDied","Data":"4b13826e640379092bec4694caca4597d2cf711ee8d2942282160b728d22acf6"} Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.557096 4756 scope.go:117] "RemoveContainer" containerID="ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.557240 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-27x92" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.580869 4756 scope.go:117] "RemoveContainer" containerID="46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.601510 4756 scope.go:117] "RemoveContainer" containerID="2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.615356 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-27x92"] Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.623760 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-27x92"] Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.650046 4756 scope.go:117] "RemoveContainer" containerID="ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3" Feb 03 10:17:14 crc kubenswrapper[4756]: E0203 10:17:14.650511 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3\": container with ID starting with ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3 not found: ID does not exist" containerID="ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.650566 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3"} err="failed to get container status \"ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3\": rpc error: code = NotFound desc = could not find container \"ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3\": container with ID starting with ee3faaf3722c85911d3fb1f9c0d7e8210e0f575003441e2ad5fa2a9706e320b3 not found: ID does not exist" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.650603 4756 scope.go:117] "RemoveContainer" containerID="46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88" Feb 03 10:17:14 crc kubenswrapper[4756]: E0203 10:17:14.650995 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88\": container with ID starting with 46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88 not found: ID does not exist" containerID="46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.651048 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88"} err="failed to get container status \"46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88\": rpc error: code = NotFound desc = could not find container \"46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88\": container with ID starting with 46b0933011e6f5f7e4137c8f165513dfb955a41989a14a3335d278f42ebf6e88 not found: ID does not exist" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.651083 4756 scope.go:117] "RemoveContainer" containerID="2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6" Feb 03 10:17:14 crc kubenswrapper[4756]: E0203 10:17:14.651386 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6\": container with ID starting with 2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6 not found: ID does not exist" containerID="2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6" Feb 03 10:17:14 crc kubenswrapper[4756]: I0203 10:17:14.651423 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6"} err="failed to get container status \"2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6\": rpc error: code = NotFound desc = could not find container \"2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6\": container with ID starting with 2cb607642003bc7b82d7f02b75b94827a215a501be8c825cdc75495b48e1e1c6 not found: ID does not exist" Feb 03 10:17:15 crc kubenswrapper[4756]: I0203 10:17:15.643041 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" path="/var/lib/kubelet/pods/ef90a0a3-2ed0-494f-9278-2bccf7ea1c42/volumes" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.052890 4756 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwzrz"] Feb 03 10:17:36 crc kubenswrapper[4756]: E0203 10:17:36.054010 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="registry-server" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.054031 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="registry-server" Feb 03 10:17:36 crc kubenswrapper[4756]: E0203 10:17:36.054052 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="extract-utilities" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.054061 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="extract-utilities" Feb 03 10:17:36 crc kubenswrapper[4756]: E0203 10:17:36.054083 4756 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="extract-content" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.054094 4756 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="extract-content" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.054386 4756 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef90a0a3-2ed0-494f-9278-2bccf7ea1c42" containerName="registry-server" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.056248 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.071924 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwzrz"] Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.172717 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tczdd\" (UniqueName: \"kubernetes.io/projected/1fede15c-1dfc-4d63-9afa-f569c4cee290-kube-api-access-tczdd\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.172812 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-utilities\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.174646 4756 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-catalog-content\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.276814 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-catalog-content\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.276881 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tczdd\" (UniqueName: \"kubernetes.io/projected/1fede15c-1dfc-4d63-9afa-f569c4cee290-kube-api-access-tczdd\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.276905 4756 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-utilities\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.277506 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-utilities\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.277779 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-catalog-content\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.315887 4756 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tczdd\" (UniqueName: \"kubernetes.io/projected/1fede15c-1dfc-4d63-9afa-f569c4cee290-kube-api-access-tczdd\") pod \"community-operators-nwzrz\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.386395 4756 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:36 crc kubenswrapper[4756]: I0203 10:17:36.991139 4756 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwzrz"] Feb 03 10:17:37 crc kubenswrapper[4756]: I0203 10:17:37.783823 4756 generic.go:334] "Generic (PLEG): container finished" podID="1fede15c-1dfc-4d63-9afa-f569c4cee290" containerID="98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10" exitCode=0 Feb 03 10:17:37 crc kubenswrapper[4756]: I0203 10:17:37.784021 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwzrz" event={"ID":"1fede15c-1dfc-4d63-9afa-f569c4cee290","Type":"ContainerDied","Data":"98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10"} Feb 03 10:17:37 crc kubenswrapper[4756]: I0203 10:17:37.784123 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwzrz" event={"ID":"1fede15c-1dfc-4d63-9afa-f569c4cee290","Type":"ContainerStarted","Data":"1da91f678039fd1f5b80dc059b5799e84cb33b8f217284613ad0d6103a298c04"} Feb 03 10:17:38 crc kubenswrapper[4756]: I0203 10:17:38.794785 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwzrz" event={"ID":"1fede15c-1dfc-4d63-9afa-f569c4cee290","Type":"ContainerStarted","Data":"6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc"} Feb 03 10:17:39 crc kubenswrapper[4756]: I0203 10:17:39.826829 4756 generic.go:334] "Generic (PLEG): container finished" podID="1fede15c-1dfc-4d63-9afa-f569c4cee290" containerID="6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc" exitCode=0 Feb 03 10:17:39 crc kubenswrapper[4756]: I0203 10:17:39.827217 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwzrz" event={"ID":"1fede15c-1dfc-4d63-9afa-f569c4cee290","Type":"ContainerDied","Data":"6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc"} Feb 03 10:17:40 crc kubenswrapper[4756]: I0203 10:17:40.837872 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwzrz" event={"ID":"1fede15c-1dfc-4d63-9afa-f569c4cee290","Type":"ContainerStarted","Data":"dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825"} Feb 03 10:17:40 crc kubenswrapper[4756]: I0203 10:17:40.858150 4756 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwzrz" podStartSLOduration=2.350081977 podStartE2EDuration="4.85812974s" podCreationTimestamp="2026-02-03 10:17:36 +0000 UTC" firstStartedPulling="2026-02-03 10:17:37.786712058 +0000 UTC m=+4048.937179443" lastFinishedPulling="2026-02-03 10:17:40.294759841 +0000 UTC m=+4051.445227206" observedRunningTime="2026-02-03 10:17:40.853298549 +0000 UTC m=+4052.003765924" watchObservedRunningTime="2026-02-03 10:17:40.85812974 +0000 UTC m=+4052.008597115" Feb 03 10:17:43 crc kubenswrapper[4756]: I0203 10:17:43.565834 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:17:43 crc kubenswrapper[4756]: I0203 10:17:43.566152 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:17:46 crc kubenswrapper[4756]: I0203 10:17:46.388041 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:46 crc kubenswrapper[4756]: I0203 10:17:46.388988 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:46 crc kubenswrapper[4756]: I0203 10:17:46.442618 4756 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:46 crc kubenswrapper[4756]: I0203 10:17:46.973738 4756 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:47 crc kubenswrapper[4756]: I0203 10:17:47.030037 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwzrz"] Feb 03 10:17:48 crc kubenswrapper[4756]: I0203 10:17:48.912878 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwzrz" podUID="1fede15c-1dfc-4d63-9afa-f569c4cee290" containerName="registry-server" containerID="cri-o://dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825" gracePeriod=2 Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.458277 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.574392 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-catalog-content\") pod \"1fede15c-1dfc-4d63-9afa-f569c4cee290\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.574532 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tczdd\" (UniqueName: \"kubernetes.io/projected/1fede15c-1dfc-4d63-9afa-f569c4cee290-kube-api-access-tczdd\") pod \"1fede15c-1dfc-4d63-9afa-f569c4cee290\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.574657 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-utilities\") pod \"1fede15c-1dfc-4d63-9afa-f569c4cee290\" (UID: \"1fede15c-1dfc-4d63-9afa-f569c4cee290\") " Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.576142 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-utilities" (OuterVolumeSpecName: "utilities") pod "1fede15c-1dfc-4d63-9afa-f569c4cee290" (UID: "1fede15c-1dfc-4d63-9afa-f569c4cee290"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.586112 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fede15c-1dfc-4d63-9afa-f569c4cee290-kube-api-access-tczdd" (OuterVolumeSpecName: "kube-api-access-tczdd") pod "1fede15c-1dfc-4d63-9afa-f569c4cee290" (UID: "1fede15c-1dfc-4d63-9afa-f569c4cee290"). InnerVolumeSpecName "kube-api-access-tczdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.636889 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fede15c-1dfc-4d63-9afa-f569c4cee290" (UID: "1fede15c-1dfc-4d63-9afa-f569c4cee290"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.677358 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tczdd\" (UniqueName: \"kubernetes.io/projected/1fede15c-1dfc-4d63-9afa-f569c4cee290-kube-api-access-tczdd\") on node \"crc\" DevicePath \"\"" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.677401 4756 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-utilities\") on node \"crc\" DevicePath \"\"" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.677414 4756 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fede15c-1dfc-4d63-9afa-f569c4cee290-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.921729 4756 generic.go:334] "Generic (PLEG): container finished" podID="1fede15c-1dfc-4d63-9afa-f569c4cee290" containerID="dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825" exitCode=0 Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.921772 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwzrz" event={"ID":"1fede15c-1dfc-4d63-9afa-f569c4cee290","Type":"ContainerDied","Data":"dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825"} Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.921800 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwzrz" event={"ID":"1fede15c-1dfc-4d63-9afa-f569c4cee290","Type":"ContainerDied","Data":"1da91f678039fd1f5b80dc059b5799e84cb33b8f217284613ad0d6103a298c04"} Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.921800 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwzrz" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.921817 4756 scope.go:117] "RemoveContainer" containerID="dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.942353 4756 scope.go:117] "RemoveContainer" containerID="6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc" Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.971484 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwzrz"] Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.980273 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwzrz"] Feb 03 10:17:49 crc kubenswrapper[4756]: I0203 10:17:49.983706 4756 scope.go:117] "RemoveContainer" containerID="98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10" Feb 03 10:17:50 crc kubenswrapper[4756]: I0203 10:17:50.039290 4756 scope.go:117] "RemoveContainer" containerID="dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825" Feb 03 10:17:50 crc kubenswrapper[4756]: E0203 10:17:50.039679 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825\": container with ID starting with dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825 not found: ID does not exist" containerID="dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825" Feb 03 10:17:50 crc kubenswrapper[4756]: I0203 10:17:50.039711 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825"} err="failed to get container status \"dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825\": rpc error: code = NotFound desc = could not find container \"dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825\": container with ID starting with dcb767a779e47f926a89be641a234685c633f1d7ceb444fe0fb0c3400a7c0825 not found: ID does not exist" Feb 03 10:17:50 crc kubenswrapper[4756]: I0203 10:17:50.039742 4756 scope.go:117] "RemoveContainer" containerID="6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc" Feb 03 10:17:50 crc kubenswrapper[4756]: E0203 10:17:50.040106 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc\": container with ID starting with 6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc not found: ID does not exist" containerID="6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc" Feb 03 10:17:50 crc kubenswrapper[4756]: I0203 10:17:50.040130 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc"} err="failed to get container status \"6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc\": rpc error: code = NotFound desc = could not find container \"6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc\": container with ID starting with 6643421f68e41f75e6a86b0a85efe93bd81cec13651f00baf57821c2dcafcadc not found: ID does not exist" Feb 03 10:17:50 crc kubenswrapper[4756]: I0203 10:17:50.040143 4756 scope.go:117] "RemoveContainer" containerID="98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10" Feb 03 10:17:50 crc kubenswrapper[4756]: E0203 10:17:50.040398 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10\": container with ID starting with 98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10 not found: ID does not exist" containerID="98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10" Feb 03 10:17:50 crc kubenswrapper[4756]: I0203 10:17:50.040508 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10"} err="failed to get container status \"98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10\": rpc error: code = NotFound desc = could not find container \"98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10\": container with ID starting with 98b9a0c4f3ec508902d053741c427005d768c50252c6f96599ce0066321cbd10 not found: ID does not exist" Feb 03 10:17:51 crc kubenswrapper[4756]: I0203 10:17:51.630009 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fede15c-1dfc-4d63-9afa-f569c4cee290" path="/var/lib/kubelet/pods/1fede15c-1dfc-4d63-9afa-f569c4cee290/volumes" Feb 03 10:18:13 crc kubenswrapper[4756]: I0203 10:18:13.566694 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:18:13 crc kubenswrapper[4756]: I0203 10:18:13.567299 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:18:13 crc kubenswrapper[4756]: I0203 10:18:13.567347 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 10:18:13 crc kubenswrapper[4756]: I0203 10:18:13.568100 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"659a44e09cf2712199a26e5ab73d9a728d8d93116f1d5994260fd77abcbbfc49"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 10:18:13 crc kubenswrapper[4756]: I0203 10:18:13.568149 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://659a44e09cf2712199a26e5ab73d9a728d8d93116f1d5994260fd77abcbbfc49" gracePeriod=600 Feb 03 10:18:14 crc kubenswrapper[4756]: I0203 10:18:14.146147 4756 generic.go:334] "Generic (PLEG): container finished" podID="c56ae5f4-06ed-41ff-84e2-093504936086" containerID="7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159" exitCode=0 Feb 03 10:18:14 crc kubenswrapper[4756]: I0203 10:18:14.146293 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gc6f/must-gather-c58nj" event={"ID":"c56ae5f4-06ed-41ff-84e2-093504936086","Type":"ContainerDied","Data":"7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159"} Feb 03 10:18:14 crc kubenswrapper[4756]: I0203 10:18:14.148157 4756 scope.go:117] "RemoveContainer" containerID="7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159" Feb 03 10:18:14 crc kubenswrapper[4756]: I0203 10:18:14.151913 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="659a44e09cf2712199a26e5ab73d9a728d8d93116f1d5994260fd77abcbbfc49" exitCode=0 Feb 03 10:18:14 crc kubenswrapper[4756]: I0203 10:18:14.151996 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"659a44e09cf2712199a26e5ab73d9a728d8d93116f1d5994260fd77abcbbfc49"} Feb 03 10:18:14 crc kubenswrapper[4756]: I0203 10:18:14.152056 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerStarted","Data":"02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32"} Feb 03 10:18:14 crc kubenswrapper[4756]: I0203 10:18:14.152083 4756 scope.go:117] "RemoveContainer" containerID="1a7e25d302b20f656b244369e68ce5963c19466e95e8e382be3737c8fb88cdf3" Feb 03 10:18:15 crc kubenswrapper[4756]: I0203 10:18:15.055695 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gc6f_must-gather-c58nj_c56ae5f4-06ed-41ff-84e2-093504936086/gather/0.log" Feb 03 10:18:25 crc kubenswrapper[4756]: I0203 10:18:25.697208 4756 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gc6f/must-gather-c58nj"] Feb 03 10:18:25 crc kubenswrapper[4756]: I0203 10:18:25.697973 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4gc6f/must-gather-c58nj" podUID="c56ae5f4-06ed-41ff-84e2-093504936086" containerName="copy" containerID="cri-o://d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c" gracePeriod=2 Feb 03 10:18:25 crc kubenswrapper[4756]: I0203 10:18:25.713652 4756 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gc6f/must-gather-c58nj"] Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.168289 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gc6f_must-gather-c58nj_c56ae5f4-06ed-41ff-84e2-093504936086/copy/0.log" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.169115 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.278878 4756 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gc6f_must-gather-c58nj_c56ae5f4-06ed-41ff-84e2-093504936086/copy/0.log" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.279168 4756 generic.go:334] "Generic (PLEG): container finished" podID="c56ae5f4-06ed-41ff-84e2-093504936086" containerID="d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c" exitCode=143 Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.279210 4756 scope.go:117] "RemoveContainer" containerID="d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.279318 4756 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gc6f/must-gather-c58nj" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.309982 4756 scope.go:117] "RemoveContainer" containerID="7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.330600 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4r77\" (UniqueName: \"kubernetes.io/projected/c56ae5f4-06ed-41ff-84e2-093504936086-kube-api-access-n4r77\") pod \"c56ae5f4-06ed-41ff-84e2-093504936086\" (UID: \"c56ae5f4-06ed-41ff-84e2-093504936086\") " Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.330715 4756 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c56ae5f4-06ed-41ff-84e2-093504936086-must-gather-output\") pod \"c56ae5f4-06ed-41ff-84e2-093504936086\" (UID: \"c56ae5f4-06ed-41ff-84e2-093504936086\") " Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.369068 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c56ae5f4-06ed-41ff-84e2-093504936086-kube-api-access-n4r77" (OuterVolumeSpecName: "kube-api-access-n4r77") pod "c56ae5f4-06ed-41ff-84e2-093504936086" (UID: "c56ae5f4-06ed-41ff-84e2-093504936086"). InnerVolumeSpecName "kube-api-access-n4r77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.371393 4756 scope.go:117] "RemoveContainer" containerID="d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c" Feb 03 10:18:26 crc kubenswrapper[4756]: E0203 10:18:26.375754 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c\": container with ID starting with d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c not found: ID does not exist" containerID="d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.375808 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c"} err="failed to get container status \"d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c\": rpc error: code = NotFound desc = could not find container \"d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c\": container with ID starting with d5b2beb6030aa190c2a95a32f7af65f9b050adf5e3592c2ca11dd0305584b12c not found: ID does not exist" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.375833 4756 scope.go:117] "RemoveContainer" containerID="7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159" Feb 03 10:18:26 crc kubenswrapper[4756]: E0203 10:18:26.376870 4756 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159\": container with ID starting with 7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159 not found: ID does not exist" containerID="7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.376927 4756 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159"} err="failed to get container status \"7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159\": rpc error: code = NotFound desc = could not find container \"7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159\": container with ID starting with 7ccd273a6d8bdb14b7084215fbd72d319df5a250dec31a0e01dcacf623202159 not found: ID does not exist" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.433505 4756 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4r77\" (UniqueName: \"kubernetes.io/projected/c56ae5f4-06ed-41ff-84e2-093504936086-kube-api-access-n4r77\") on node \"crc\" DevicePath \"\"" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.536738 4756 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c56ae5f4-06ed-41ff-84e2-093504936086-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c56ae5f4-06ed-41ff-84e2-093504936086" (UID: "c56ae5f4-06ed-41ff-84e2-093504936086"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 03 10:18:26 crc kubenswrapper[4756]: I0203 10:18:26.636892 4756 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c56ae5f4-06ed-41ff-84e2-093504936086-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 03 10:18:27 crc kubenswrapper[4756]: I0203 10:18:27.623824 4756 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c56ae5f4-06ed-41ff-84e2-093504936086" path="/var/lib/kubelet/pods/c56ae5f4-06ed-41ff-84e2-093504936086/volumes" Feb 03 10:19:32 crc kubenswrapper[4756]: I0203 10:19:32.475728 4756 scope.go:117] "RemoveContainer" containerID="e3155261dcaeda50af721ff22b566956e4c5e4cb8ff6cf268aa7b87954ef6aa2" Feb 03 10:20:13 crc kubenswrapper[4756]: I0203 10:20:13.565567 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:20:13 crc kubenswrapper[4756]: I0203 10:20:13.566049 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:20:32 crc kubenswrapper[4756]: I0203 10:20:32.526211 4756 scope.go:117] "RemoveContainer" containerID="f1f9e62bf3055aa2e82a1b674d32813a3265948a791c64bb23976f381eaa2dfa" Feb 03 10:20:43 crc kubenswrapper[4756]: I0203 10:20:43.566317 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:20:43 crc kubenswrapper[4756]: I0203 10:20:43.566935 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.566146 4756 patch_prober.go:28] interesting pod/machine-config-daemon-c9rn9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.567514 4756 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.567593 4756 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.568358 4756 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32"} pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.568415 4756 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" containerName="machine-config-daemon" containerID="cri-o://02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" gracePeriod=600 Feb 03 10:21:13 crc kubenswrapper[4756]: E0203 10:21:13.718604 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.857526 4756 generic.go:334] "Generic (PLEG): container finished" podID="748779a5-a5e9-4451-839c-805686b764c5" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" exitCode=0 Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.857585 4756 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" event={"ID":"748779a5-a5e9-4451-839c-805686b764c5","Type":"ContainerDied","Data":"02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32"} Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.857626 4756 scope.go:117] "RemoveContainer" containerID="659a44e09cf2712199a26e5ab73d9a728d8d93116f1d5994260fd77abcbbfc49" Feb 03 10:21:13 crc kubenswrapper[4756]: I0203 10:21:13.858401 4756 scope.go:117] "RemoveContainer" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" Feb 03 10:21:13 crc kubenswrapper[4756]: E0203 10:21:13.858752 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:21:26 crc kubenswrapper[4756]: I0203 10:21:26.614824 4756 scope.go:117] "RemoveContainer" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" Feb 03 10:21:26 crc kubenswrapper[4756]: E0203 10:21:26.615634 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:21:40 crc kubenswrapper[4756]: I0203 10:21:40.614273 4756 scope.go:117] "RemoveContainer" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" Feb 03 10:21:40 crc kubenswrapper[4756]: E0203 10:21:40.615194 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:21:51 crc kubenswrapper[4756]: I0203 10:21:51.615819 4756 scope.go:117] "RemoveContainer" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" Feb 03 10:21:51 crc kubenswrapper[4756]: E0203 10:21:51.617336 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:22:03 crc kubenswrapper[4756]: I0203 10:22:03.614946 4756 scope.go:117] "RemoveContainer" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" Feb 03 10:22:03 crc kubenswrapper[4756]: E0203 10:22:03.615984 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:22:14 crc kubenswrapper[4756]: I0203 10:22:14.615001 4756 scope.go:117] "RemoveContainer" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" Feb 03 10:22:14 crc kubenswrapper[4756]: E0203 10:22:14.615981 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:22:27 crc kubenswrapper[4756]: I0203 10:22:27.615372 4756 scope.go:117] "RemoveContainer" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" Feb 03 10:22:27 crc kubenswrapper[4756]: E0203 10:22:27.616173 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" Feb 03 10:22:40 crc kubenswrapper[4756]: I0203 10:22:40.614336 4756 scope.go:117] "RemoveContainer" containerID="02de0bf63cda714d501cd3073f007a0d62a673a74bad4bad5b3723eb64f3be32" Feb 03 10:22:40 crc kubenswrapper[4756]: E0203 10:22:40.615373 4756 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-c9rn9_openshift-machine-config-operator(748779a5-a5e9-4451-839c-805686b764c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-c9rn9" podUID="748779a5-a5e9-4451-839c-805686b764c5" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515140346203024443 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015140346204017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015140335254016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015140335255015460 5ustar corecore